[ / / / / / / / / / / / / / ] [ dir / ausneets / had / htg / leftpol / radcorp / strek / sw / thestorm ]

/tech/ - Technology

Winner of the 24rd Attention-Hungry Games
/kemono/ - A match made in heaven

Comment *
* = required field[▶ Show post options & limits]
Confused? See the FAQ.
Show oekaki applet
(replaces files and can be used instead)
Password (For file and post deletion.)

Allowed file types:jpg, jpeg, gif, png, webm, mp4, pdf
Max filesize is 16 MB.
Max image dimensions are 15000 x 15000.
You may upload 3 per post.

File: e001175adc3c5fc⋯.png (87.56 KB, 1092x512, 273:128, ipfsthread.png)

File: 957095eb5bec93f⋯.png (140.19 KB, 512x512, 1:1, ipfs-logo-icewithbg-512.png)

File: 1666469e81f661b⋯.webm (3.94 MB, 1920x1040, 24:13, ipfs webm 1.webm)



0.4.10 - 2017-06-27

>Ipfs 0.4.10 is a patch release that contains several exciting new features, bugfixes and general improvements. Including new commands, easier corruption recovery, and a generally cleaner codebase.


>Add support for specifying the hash function in ipfs add

>Implement ipfs key {rm, rename}

>Implement ipfs shutdown command

>Implement ipfs pin update

>Implement ipfs pin verify

>Implemented experimental p2p commands

0.4.9 - 2017-04-30

>Ipfs 0.4.9 is a maintenance release that contains several useful bugfixes and improvements. Notably, ipfs add has gained the ability to select which CID version will be output.


>Add support for using CidV1 in 'ipfs add'

tl;dr for Beginners

>decentralized P2P network

>like torrenting, but instead of getting a .torrent file or magnet link that shares a pre-set group of files, you get a hash of the files which is searched for in the network and served automatically

>you can add files to the entire network with one line in the CLI or a drag-and-drop into the web interface

>HTTP gateways let you download any hash through your browser without running IPFS

>can stream video files in mpv or VLC (though it's not recommended unless the file has a lot of seeds)

How it Works

When you add a file, the files are cryptographically hashed and a merkle tree is created. These hashes are announced by the IPFS client to the nodes in the network. (The IPFS team often describes the network as a "Merkle forest.") Any user can request one of these hashes and the nodes set up peer connections automatically. If two users share the same file then both of them can seed it to a third person requesting the hash, as opposed to .torrent files/magnets which require both seeders use the same file.


>Is it safe?

It's about as safe as a torrent right now, ignoring the relative obscurity bonus. They are working on integration with TOR and I2P. Check out libp2p if you're curious.

>Is it fast?

Finding a seeder can take anywhere from a few seconds to a few minutes. It's slowly improving but still requires a fair bit of optimization work. Once the download starts, it's as fast as the peers can offer, just like a torrent.

>Is it a meme?

You be the judge.

It has implementations in Go (meant for desktop integration) and Javascript (meant for browser/server integration) in active development that are functional right now, it has a bunch of side projects that build on it, and it divides important parts of its development (IPLD, libp2p, etc) into separate projects that allow for drop-in support for many existing technologies.

On the other hand, it's still alpha software with a small userbase and has poor network performance.

Websites of interest


Official IPFS HTTP gateway. Slap this in front of a hash and it will download a file from the network. Be warned that this gateway is slower than using the client and accepts DMCAs.


Pomf clone that utilizes IPFS. Currently 10MB limit.

Also hosts a gateway at gateway.glop.me which doesn't have any DMCA requests as far as I can tell.


IPFS index, has some links (add ipfs.io/ before to access without installing IPFS)


>It's about as safe as a torrent right now

no it's not. bittorrent uses sha1 which was shattered and has been deprecated decades ago.


where's the release that doesn't crash routers by holding 1500 connections open



>his router cant handle >10^5 connections




I think they've fixed it.



>running 100% of your internet traffic through post-2013 pozzed hardware with proprietary firmware


If gateways make IPFS accessible to everyone, why isn't it more widely used?



because nobody uses it.



Client is badly optimized right now, it takes loads of ram and CPU for what bittorrent clients can do without breaking a sweat, they haven't worked on optimization since the protocol is constantly changing.



>the protocol is constantly changing

this is also why nobody uses it

stop breaking fucking links every 2 weeks


Why did you make a new thread? The old one was fine >>771999


>stop breaking fucking links every 2 weeks

You have no clue what you're talking about. The only non-backwards compatible change they made was in April 2016 when they released 0.4.0. Since then all changes have been backwards compatible with new releases.



It had no image.



Launch it with

ipfs daemon --routing=dhtclient

to reduce the amount of connections it uses.

Additionally, some progress has been made limiting the extent of the problem, though the issue of closing connections doesn't appear to be touched yet.

Issues to watch on this subject:




If you're going to make a new thread, at least post some updates.

js-ipfs 0.26 Released


>Here are some of the highlights for this new js-ipfs release. There were plenty more bug fixes, tiny performance improvements, doc improvements and others all across the js-ipfs module ecosystem. A really BIG THANK YOU to everyone that has been contributing with code, tests, examples and also bug reports! They help us identify situations that we miss without tests.

New InterPlanetary Infrastructure

>You might have noticed some hiccups a couple of weeks ago. That was due to a revamp and improvement in our infrastructure that separated Bootstraper nodes from Gateway nodes. We’ve now fixed that by ensuring that a js-ipfs node connects to all of them. More nodes on https://github.com/ipfs/js-ipfs/issues/973 and https://github.com/ipfs/js-ipfs/pull/975. Thanks @lgierth for improving IPFS infra and for setting up all of those DNS websockets endpoints for js-ipfs to connect to :)

Now js-ipfs packs the IPFS Gateway as well

>You read it right! Now, js-ipfs packs the IPFS Gateway and launches it when you boot a daemon (jsipfs daemon). With this, you can use js-ipfs to access content in the browser just like you use to do in go-ipfs or use js-ipfs as a complete solution to add content in the network and preview it without leaving JS land. It is great for tooling. This was an awesome contribution from @ya7ya and @harshjv who spent a lot of time adjusting and iterating on the implementation to make sure it would fit with the structure of js-ipfs, 👏🏽👏🏽👏🏽👏🏽.

Huge performance and memory improvement

>With reports such as https://github.com/ipfs/js-ipfs/issues/952, we started investigating what were the actual culprits for such memory waste that would lead the browser to crash. It turns out that there were two and we got one fixed. The two were:

>>browserify-aes - @dignifiedquire identified that there were a lot of Buffers being allocated in browserify-aes, the AES shim we use in the browser (this was only a issue in the browser) and promptly came with a fix https://github.com/crypto-browserify/browserify-aes/pull/48 👏🏽👏🏽👏🏽👏🏽

>>WebRTC - WebRTC is really cpu+mem hungry and our combination of opening multiple connections without limits + the constant switch between transport and routing at the main thread, leads to some undesirable situations where the browser simply crashes for so much thrashing. We are actively working on this with Connection Closing.

>That said, situations such as https://github.com/ipfs/js-ipfs/issues/952 are now fixed. Happy file browser sharing! :)

Now git is also one of the IPLD supported formats by js-ipfs

>Now js-ipfs supports ipld-git! This is huge, it means that you can traverse through git objects using the same DAG API that you use for Ethereum, dag-pb and dag-cbor. This feature came in with an example, go check out how to traverse a git repo. 👏🏽👏🏽 to @magik6k for shipping this in record time.

The libp2p-webrtc-star multiaddrs have been fixed

>@diasdavid (me) and @lgierth had a good convo and reviewed a bunch of stuff over Coffee ☕️, it was great!. During that chat, we figured that libp2p-webrtc-star multiaddrs have been implemented incorrectly and figured out the migration path to the correct version.

>You can learn more what this endeavour involved here https://github.com/ipfs/js-ipfs/issues/981. Essentially, there are no more /libp2p-webrtc-star/dns4/star-signal.cloud.ipfs.team/wss, instead we use /dns4/star-signal.cloud.ipfs.team/wss/p2p-webrtc-star which signals the proper encapsulation you expect from a multiaddr.

New example showing how to stream video using hls.js

>@moshisushi developed a video streamer on top of js-ipfs and shared an example with us. You can now find that example as part of the examples set in this repo. Check https://github.com/ipfs/js-ipfs/tree/master/examples/browser-video-streaming, it is super cool 👏🏽👏🏽👏🏽👏🏽.

>HLS (Apple’s HTTP Live Streaming) is one of the several protocols currently available for adaptive bitrate streaming.

webcrypto-ossl was removed from the dependency tree

>We’ve purged webcrypto-ossl from the dependency tree. It was only used to generate RSA keys faster (significantly faster) but at the same time, it caused a lot of hurdles to being a native dependency.

>There is an open issue on the Node.js project to expose the RSA key generation primitive, if you have a use case for it please do share it in that thread. This would enable js-ipfs to use the native crypto module, have the same (or better) perf of webcrypto-ossl and not have to deal with an external native dependency.

PubSub tutorial published

>@pgte published an amazing tutorial on how to use PubSub with js-ipfs and in the browser! Read it on the IPFS Blog https://blog.ipfs.io/29-js-ipfs-pubsub.


do ipfs still use the ipfs (((package))) app that

throws away the cryptographic integrity of git?


do ipfs still use the ipfs (((package))) app that

throws away the cryptographic integrity of git?




NixOS will have IPFS as well WEW


this is nice but i can do without the pedoshit links


Call us when it can use both i2p/tor. Until then, it's inferior to bittorrent for filesharing. Why? Because the only client availables (https://github.com/Agorise/c-ipfs isn't ready enough) are in GC using retarded langages.



They're working on it. c-ipfs seems promising as well.



What pedo links?



Turn on IPv6 you dumb nigger. NAT is cancer.


Does IPFS need the equivalent of TOR firefox?


My quick IPFS indexer seems to be working good, will try to make database available over IPFS if it keeps working good. Could anyone who knows how ES works download the 200gb of dumps from ipfs-search and run `grep --only-matching -P "[0-9A-Za-z]{46}"` on it? They're compressed/encrypted somehow.


Nah, it's non-anonymous right now anyway, you could just install a new chromium/whatever and limit it to localhost:8080 if you really want to.


Can the anon behind this comment on whether js-ipfs 0.26 is enough to get it started?




I'd rather put my hopes into the actually-existing ipfs ib made by another anon here over anime-pic-one-commit over there.


>ipfs add 60MB in three files

>router reboots

This is 16chan-tier software



Now is not the time for optimization, that comes later.

Also, see >>793346


File: 0bba07533f7cb3f⋯.png (536.4 KB, 1276x531, 1276:531, 1501285882315.png)

File: 1aa9ee532fabc18⋯.png (340.73 KB, 1214x595, 1214:595, 1505575841282.png)

"The age of men will return;

and they're not gonna get their computer-grid, self-driving car, nano-tech, panopticon in place fast enough..."




Why does IPFS idle at like 25% CPU usage and high ram usage? Its not even seeding often or downloading anything.



How much later? They've known this shit is unusable for two years.


File: d9a40bd09bcdbff⋯.png (197.19 KB, 842x848, 421:424, 1494016389196.png)

what does it do better than zeronet



Better spec, better design, and better future-proofed.

Problem is, right now it still send a fuckton of diagnostics information and is poorly optimized.

They're making progress, but it's not fast enough.



It's not a honeypot where every post you make on any site is tracked globally and anyone can insert arbitrary javascript (automatically run by all clients) (which must be enabled completely at all time to view sites in the first place) so long as any content of any type is allowed for upload in any page where that content would appear, for one.



bittorrent is NOT secure over TOR



It's just as secure as any other protocol. Some clients might potentially leak your IP, but there is nothing inherently insecure about the protocol.


File: c8a0ec27ffbb3b4⋯.png (143.55 KB, 445x385, 89:77, 1429497660865.png)


>anyone can insert arbitrary javascript (automatically run by all clients)

That sounds bad. Are you sure? It sounds so bad that I don't know if I believe you.



Try it yourself. The steps are as follows:

- go to an arbitrary site

- upload whatever they allow, content is irrelevant

- you will find a local file that was created with the content you uploaded

- edit it to insert arbitrary content

simple as that, bypasses all sanitation attempts, etc.

There has been proofs of concepts on 0chan in the very early days of zeronet, and this has never been addressed. The zeronet devs seem to not care about any of the components that make this possible.



sasuga redditware


This is all well and good but the problem with the internet today is that it relies on a centralized infrastructure to gain access to which one has to pay fees to people with the power to cut one off completely if not control the content being served.





>>Is it a meme?

>You be the judge.

>It has implementations in Go (meant for desktop integration) and Javascript (meant for browser/server integration) in active development that are functional right now, it has a bunch of side projects that build on it, and it divides important parts of its development (IPLD, libp2p, etc) into separate projects that allow for drop-in support for many existing technologies.

i have judged



Try upgrading from a Pentium 3.


Also centralized web servers were conceived because of a very important flaw with peer to peer infrastructure

>guy who owns file turns off machine

>file is gone



but you've got that wrong, a centralized server shuts down no one can ever access that file.



Did you ponder a bit before typing such a sloppy mess? Those centralized cluster of web servers can still represent a peer in a p2p infrastructure. Nothing is forbidding such a cluster from joining the swarm and sharing files. Compared to a distributed web, centralization offers very little benefits in return and is only still around because it offers more control (to the owners of the files and servers).



This is literally the same problem with web servers. Even your VPS is on an actual server. So if you're saying P2P (and self-hosting) has flaws well

>If the guy who owns AWS doesn't like you, no one can ever access that file.





none of you are wrong but you've all missed his point



What point, that you can't get a file if the only peer in the world that has it turns off his PC? Well no shit captain obvious, but that isn't a problem with the p2p infrastructure, that's a problem of people not giving enough of a shit to seed that file. Maybe Filecoin is a better solution to that, but then again who knows what will happen in the future.


File: 06a287e663c5599⋯.png (54.21 KB, 557x380, 557:380, Screenshot from 2017-09-19….png)

File: 822c3d9bb8c8cb8⋯.jpg (366.2 KB, 1024x768, 4:3, a-programmer-describes-how….jpg)



jesus christ OP I just wanna download Initial D and DBZ and you're linking to CP.



you talk too much man


File: 8c1d4c2f0af62f8⋯.jpg (38.63 KB, 374x374, 1:1, D8CRtMS.jpg)

can anyone explain how i can use p2p in a website. i have an idea of how i want to use it for a chan and other ways but i dont understand how i would go about it.



are you asking for an entry level explanation of how it works?



no, how would you apply p2p to a traditional website.

also what i dont like about ipfs is that it's almost impossible to have any privacy or remove content.



>no, how would you apply p2p to a traditional website.

decentralized hosting, I'm not sure I understand the question?



im talking about the code, how would you apply p2p.



that depends on the type of site anon, come on



well, let me make a mockup.


File: a0523f9e866e2de⋯.png (510.18 KB, 2354x946, 107:43, 643d.png)


I want to make a chan where users can create their own boards on the host site, and every thread is hosted by the users though p2p, the more users the more relevance and faster the thread loads for everyone in the thread. after a certain number of posts the thread will be removed and flushed out of everyone's computer.

the features will be unlimited file /video and reasonably lengthy text sizes.


File: d552e926579b143⋯.png (33.33 KB, 1383x163, 1383:163, zn.png)



that's almost exactly the same as IPFS chan



sorry for asking to be spoonfed, but do you mind showing me a bug about this? I can't find it.

Based on what I can tell, you can't just arbitrarily change 0chan to post whatever file you want.

This isn't related to ipfs, so I'll sage



Anon's just havin' a giggle, go ahead and click it.



Unoptimized, probably the DHT.


They're working on other stuff right now, they got a lot of money from filecoin ICO so we should be seeing some progress pretty soon. There's also a C implementation in the works.


IPFS can run on any transport (hyperboria), which can run over any link (ronja)


You cache it automatically when you download it.


Tor integration is in the works for privacy


>create keypair

>use this as board name

>sign admin keys

>they sign lists of posts to filter

>posts which do not follow specifications (eg too long text) are filtered by clients

>threads which are too old are filtered by clients



You're mostly just describing smugboard >>785171, it's very similar to what you're proposing.



It's not a bug. It's how it is designed. The poster can arbitrarily change the content of what they have posted. This includes changing the media type, and is simply a matter of editing the content that is stored locally. When someone requests the file, because you are the poster of the file, your doctored copy is distributed because the content you "upload" (e.g. text posts, or actual document attachments, etc.) are handled in this way.

You can even see the instructions on "how to modify a zeronet site" here:


as comments you post to a site are not handled in any special way compared to anything else. It's also why you need an ID to post anything and why your ID can be used to track anything you say across all sites by a simple grep: to enable modifying the content (which is not differentiable from a site, up to a point) by signing a more recent copy of the content.



wew, its worse than I thought


File: 945a274358a52c1⋯.png (83.63 KB, 300x300, 1:1, 1397153602672.png)


Why aren't files hashed? IPFS gets this right, why is there no network-level guarantee that files haven't been altered?



but Tor works easily on a P3, why is IPFS special?



IPFS is still in alpha (not optimized yet) and has the overhead of a complete p2p system (routing). Tor is much simpler to implement since every peer is not a contributing node. A large number of peers connect to a limited amount fast nodes. In IPFS every peer is also a node. This is the difference betwen Tor's decentralized network approach and IPFS's distributed network approach.



IPFS uses a distributed naming system (ipns) to point to the latest version as well as static pointers (ipfs-based addresses) to point to specific files. This is to enable tracking the latest version (i.e. enable the ability to update content) while still giving the guarantees there was no tempering by the controller. Zeronet doesn't seem to care at all about such possibilities: all that matters, only about the ability to update the content. Similarly, the zeronet folks don't give a shit about security (for the longest time (that might still be the case) they had been running with very old versions of various libs, including crypto libs, with unaddressed CVEs, for example). You can just say "their threat model is different" but at this point they disregard secops 101.


Release candidates are out for go-ipfs 0.4.11. If you want to try them out, check out the download page: https://dist.ipfs.io/go-ipfs

If you have troubles with IPFS using way too much bandwidth (especially during add), memory leaks, or running out of file descriptors, you may want to make the jump as soon as possible. This version includes prototypes for a lot of new features designed to improve performance all around.



So if I use something like IFPS to host a website, does that mean that I don't have do fuck with things like domain name registration?



Technically yes, but in practice you will still need a way to register a user-friendly name because people can't recall localhost:8080/ipns/LEuori324n2klAJFieow. But there's a way to add a friendly name in the ipns system (google around, I don't recall the correct method), which allows people to use localhost:8080/ipns/your.address.name instead, so that's an option. Other than that, all kinds of systems can leverage the likes of namecoin if you're so inclined.



Eh, I actually prefer the hash method. Keeps things a little more comfy.

>In the future, we may have ipns entries work as a git commit chain, with each successive entry pointing back in time to other values.

That's really fucking cool though.




Here's a list of all DNS alternatives that IPFS team can use, my guess is that they will use Filecoin considering that it belongs to them.



The method is to register with any normal DNS method a TXT record with content: dnslink="/ipns/<hash>" and it will work. So it's actually relying on the external system.


I thought filecoin was just an incentive to store other people's files?


File: 7dae464e0cfa70c⋯.jpg (73.49 KB, 768x780, 64:65, shocking truth.jpg)

Can I limit the amount of space that IPFS uses or if I download and start running it will it just fill up my hard drive indefinitely?



>I thought filecoin was just an incentive to store other people's files?

It is. I think they're going to recommend using ethereum domains as IPFS has plans to be deeply integrated with it.


>Can I limit the amount of space that IPFS uses

IPFS doesn't download random things to your computer. It caches everything you view but by default it's capped at 10GB.



By default IPFS does not fetch anything on its own, it only will retain the data you manually added via browsing or manual adding.

If you want you can run the daemon like this `ipfs daemon --enable-gc` which will read your config for 2 values, 1 is a timer and the other is storage. By default I think they're 1 hour and 10GBs, that means a gabrage collection routine would run either when you hit 10GB's of garbage or 1 hour has passed. What it considers garbage is anything that's not "pinned", if you don't want something to be treated like garbage you pin it.

Someone made an issue recently that I agree with, there should be an option for a minimum amount of data to keep, right now garbage collection deletes ALL garbage, but it would be nice if you could set it to keep xGB's worth of non-pinned content at any one time.



File: eceb197dec09787⋯.png (870.38 KB, 820x650, 82:65, manga_13set (1).png)


All 13 of the current "Manga Guides" series, in various formats.


File: 20b5e558d8cfc0c⋯.jpg (121.37 KB, 900x1200, 3:4, water between azn tits.jpg)


My mixtape.


Good music with a good video to go with it


Holy Nonsense

Also why does

32.00 MB / 54.44 MB [=======================================================================================================================>------------------------------------------------------------------------------------] 58.79% 0s20:13:33.012 ERROR commands/h: open /home/anon/.ipfs/blocks/GY/put-460657004: too many open files client.go:247
Error: open /home/anon/.ipfs/blocks/GY/put-460657004: too many open files

keep happening? Each of these I had to try adding several times.



Have you upgraded to 0.4.11 yet?


I have a question about implementation

Each file is divided into chunks, which are then hashed. These hashed chunks form the leaves of the merkle tree, which have parents that are identified by HASH( HASH( left-child ) + HASH( right-child )). This continues until we reach the root node, the merkle root, whose hash uniquely identifies the file.

To give someone else the file, from computer S to computer T, S gives T the list of leaves, and the merkle root. As I understand it, this is basically what a bittorent magnet link does as well (along with tracker and other metadata). We know the leaves actually compose the merkle root, by simply building the tree from its leaves, and verifying the new merkle root is the same as the provided one.

Computer T then ask around if anyone else has the content of the leaves (by querying for the leaf-hash), and verifies the content by hashing it upon download-completion. Once it has everything (and verifies), it simply compiles the parts into the

Assuming there is nothing wrong with my understanding above, I have a few questions:

How do we know the merkle root actually identifies the file we meant to get? ie if someone hits an IPNS endpoint, and an attacker intercepts and returns a malicious merkle root + leaves, now what? Is there anything to do about this or is this just a case of don't trust sites you don't know

When computer T starts requesting for leaf-content, is it requesting by querying on the hash of a leaf, or the merkle root? Bittorent only requests parts from users that have the full file, which comes from the latter. If you request by the leaf-hash instead, I'm imagining that the less-unique parts (like say a chunk of the file composed entirely of NULL bytes) could come from ANY file source, regardless if that user actually has the file you're looking for.

And extending that, with some infinite number of files stored globally, it would be possible to download files with a leaf-list that NO ONE actually has; each leaf being found in some other file; composed in some particular fashion to create the requested file.


Can you use IPFS in combination with a tor bridge with obfs encrypting files and still transfer files? If this worked would the person receiving the data still see your public IP?



>How do we know the merkle root actually identifies the file we meant to get?

So you mean how is the data verified to be correct once the client receives it. inb4 is isn't verified



>When computer T starts requesting for leaf-content, is it requesting by querying on the hash of a leaf, or the merkle root?

On the leaf. Each 256k block has its own DHT entry (hence why it's known to be so chatty). This also means that if you have a file and change one byte in it then most of the file will be deduped by IPFS if you readd it.


>if someone hits an IPNS endpoint, and an attacker intercepts and returns a malicious merkle root + leaves, now what?

My understanding is that the IPNS entries are signed by your public key, so that's not an issue. There is a problem where a malicious node could return an old entry, but that's the reason each entry is stored in a fuck-ton of DHT nodes. Which is also the reason it takes so long to resolve names, it doesn't just take the first resolution it can.



>On the leaf.

So what I suggested then, that a file no one has could be generated by the network given a list of leaves, by retrieving them from other files, would hold then? I suppose though that's not anything special, except that the granularity of chunks is bigger than say, 1 bit. But to confirm my understanding, is this true?



With bitswap, it DOES download random things, kinda. You swap fragments among peers from random content.



It's hash-based addressing: if the chunks that make up the file exist in other files, they are exactly as valid in the requested file as it is in that other file. That is, request the hash and the provenance is a meaningless concept: you can think of it as two completely different kinds of data (the actual chunks, and the file descriptors which are merkel graphs)



> ie if someone hits an IPNS endpoint, and an attacker intercepts and returns a malicious merkle root + leaves, now what?

IPNS is still handled via DNS, so short of someone pwning the authoritative nameserver for a domain, you're looking at a hijacked local resolver, which you can defeat via VPN.


File: ef73c99feafef30⋯.jpg (57.52 KB, 261x287, 261:287, 1385077398214.jpg)



I am requesting that people please prefix their hashes with"/ipfs/" when posting, so that the browser addon detects them and anchors them, this way people with it can just click on them.

Like this





updated my porn folder again




direct link since ipns is buggy



alright my first try at this, it's Lovecrafts "beyond the wall of sleep"



>merkel tree

/pol/ shoop incoming




Nobody is interested in the contents of your spank folder, you degenerate.



but someone might be


File: 047afe01d717493⋯.jpg (38.18 KB, 640x649, 640:649, CRjSbwjVEAAuScv.jpg large.jpg)


And here the necronomicon


could someone tell me if it works alright?



You can check yourself by accessing a file through the gateway, i.e. https://ipfs.io/ipfs/QmT45tFQo5DJ8m7VShLPecsTsGy1aSKBq4Pww8MfaMppK6

If you can see it there, everyone can find it.



>ipns is buggy

How do you mean?



ipns forgets what it's linked to after 12 hours and is extremely slow



>Right now if you are Catalonian you can go and look up where to vote for the independence of your country via an encrypted and fully distributed web running on experimental, unfinished free and open source software maintained by a global network of volunteers, which is the only way to do it since the government censored the original website and is sending the cops after everyone who tries to mirror it normally. The future is now, motherfuckers.



RIght near the top of page linked by OP is the following:

gathered links. mostly from IPFS generals on 8ch/tech/.

* anon's qt ipfs page (this site)


* the best cp archive i've ever seen


I ain't clicking it, regardless of what it may hold. Probably cartoon ponies teaching how to circuit-probe.



>And here the necronomicon

>could someone tell me if it works alright?

No one can *tell* you that magick works, Anon. You have to say the incantation yourself. If you see results that seem magical to *you* then the magick has worked for *your* belief. Magick is not science.



Disclaimer: I think this is all right but I'm not sure since things are changing rapidly, correct me if I'm wrong

Nodes don't forget, the record expires, by default your daemon is supposed to re-publish anything you published every 12 hours and records expire every 24


Just make sure you actually publish after starting the daemon and again anytime the daemon restarts. I think this will be automated later and other peers will keep the record alive but it's not like that yet.


It's Rick Astley's Never Gonna Give You Up


Are you still hosting? I managed to get only 25MB's.



If magic only exists based on personal perception then magic cannot affect other people in any way, since their perception differs from yours.



yeah I forgot to turn it on, try again.


IPFS v0.4.11 is out

This looks like a large update. Everybody should upgrade.

>You will now be able to configure ipfs datastore to use flatfs, leveldb, badger, an in-memory datastore, and more to suit your needs

>The concept of 'Bitswap Sessions' allows bitswap to associate requests for different blocks to the same underlying session, and from that infer better ways of requesting that data

>As nodes update to this and future versions, expect to see idle bandwidth usage on the ipfs network go down noticeably

>Users who previously received "too many open files" errors should see this much less often in 0.4.1

>A memory leak in the DHT was identified and fixed and now memory usage appears to be stable over time

>In an effort to provide a more ubiquitous p2p mesh, we have implemented a relay mechanism that allows willing peers to relay traffic for other peers who might not otherwise be able to communicate with each other

>we have come up with a system that allows users to install plugins to the vanilla ipfs daemon that augment its capabilities (git and ethereum plugins are currently available)

>In the future, we will be adding plugins for other things like datastore backends and specialized libp2p network transports

>we've switched the ipfs/go-ipfs docker image from a musl base to a glibc base




>going from muslc to botnetc

Hmm. As long as you are compiling it yourself still......



2 weeks old my man.



>It's about as safe as a torrent right now


>no it's not. bittorrent uses sha1 which was shattered and has been deprecated decades ago.

Data on the attack. https://security.googleblog.com/2017/02/announcing-first-sha1-collision.html

For the time being, it's still not considered a realistic scenario because of the amount of effort involved, even if it's a possible one. Either way, they've added SHA256 to the BitTorrent v2 specifications.




>Either way, they've added SHA256 to the BitTorrent v2 specifications.

Fucking finally.



It was released the 27th. You're thinking of the release candidate.



>not using SHA2-512/256 or BLAKE2b-256

That was either a retarded decision or a (((perfectly planned))) decision. Why would they slow down the entire network to use an algorithm that isn't even recommended anymore due to the prevalent security threats? And don't give me any muh hardware acceleration crap. The only x86 CPUs that have hw acceleration are newer Intel Atoms and AND Zen which count for a minuscule percent of torrent users.



It was (((perfectly planned))) considering the (((endurance international group))) owns bittorrent now.



>And don't give me any muh hardware acceleration crap.

That's exactly the excuses I was reading in the Issue thread on the subject. It sounded stupid to me too.

>not using SHA2-512/256 or BLAKE2b-256

BitTorrent is an open standard. It's up to the community to implement. If the community got together and decided to implement a modified protocol, the recommended specs would necessarily have to change to reflect this if enough people were about it. That or we could fork and call it ButtTorrent, since we're being stubborn butts about it and everyone just calls it torrenting anyway.


File: 0a3de96b2eb93f9⋯.jpg (297.92 KB, 1224x1632, 3:4, J7NMPNR.jpg)

Would IPFS be a great choice for those who're looking for a way to have a AI back itself up to prevent another Tay?



Those titles were a joke, nigger


Do you think it will work in the outer space?




how would you prevent another tay?



>how would you prevent a large corporation from deleting its chatbot after we ask it to repeat 14/88 memes?

I don't think there's a solution to that other than do it yourself. If there's source code, you can rebuild it; if there isn't, tough luck. I don't see what IPFS has to do with this, other than maybe helping to host it or its databases.


To whomever posted QmVuqQudeX8dhPDL8SPZbngvBXHxHWiPPoYLGgBudM1LR5 and QmT45tFQo5DJ8m7VShLPecsTsGy1aSKBq4Pww8MfaMppK6 these stop part-way. Please continue to seed them.


File: 64b660b44a845f0⋯.png (180.44 KB, 500x500, 1:1, 1263746040872.png)

File: 3bd208a3981f553⋯.gif (2.86 MB, 480x480, 1:1, water_in_space.gif)

File: f7edc59750a823e⋯.png (99.5 KB, 958x574, 479:287, Puyo Mac.png)


Not the original poster but I have a copy of those but I'm currently migrating from flatfs to badger, it's going to take a few hours because I have hundreds of GBs of shit in here and slow ass hard drives. My node should come online in like 20 hours or so since I just started.


What data should we send to/from space? I'm willing to send puzzle games and cute girls if they send me pictures of the moon and videos of them doing things in low gravity.




Okay I'm up now, badger seems fast as heck compared to flatfs.

I made a mistake earlier, I only have 25% of /ipfs/QmT45tFQo5DJ8m7VShLPecsTsGy1aSKBq4Pww8MfaMppK6

I knew I had 100% of /ipfs/QmVuqQudeX8dhPDL8SPZbngvBXHxHWiPPoYLGgBudM1LR5 though

OP should come online so I can finish the mirror.



Are there any benchmarks on datastore filesystems yet?



The benchmarks for the underlying systems should be representative enough. There are plenty online of people comparing leveldb with other things, badger itself competes with an improved version of leveldb(RocksDB) and the devs posted a lot of information about both of them in their post, they also have comparison results.


There's probably no benchmark for the flatfs scheme the ipfs team made, I don't think it would make sense to bench either since the underlying filesystem and OS have a major impact on it, since it's just flat files in directories.

Reminder that the interface for datastores is in place and currently being worked on, so you can implement the interface and use whichever datastore you want. Most of (if not all of) the filestore(no-copy adds) work was done by a third party.



I suppose badger is a no brainer considering they're making it the default next patch, as far as I can tell.



It seems objectively better than leveldb in every way, personally it's been much better than leveldb and flatfs on my machine, like noticeably so in terms of speed, hashes went from taking about 2 seconds to be fetched on the public gateway, to instant, and this is with hundreds of gigabytes of things in my datastore.



Surprising how much overhead there is in flatfs just from storing everything as 256kiB files. Makes me wonder how much faster git would be with an sqlite backend.


bump for interest



What are you interested in?



they meant usury



>thinking anyone will use this

>thinking it won't get outlawed



Many people do use this without even knowing, dumb fuck. Get out, candy ass piss ant; Inform yourself before speaking. This isn't some inconvenient technology. On the contrary, very fast and improves speed over regular internet usage. This isn't even addressing the other benefits. BTFO.



Whoa, calm down there sis.


Outlawed or not it would be hard to prevent people from using it. The development team is coming up with strategies to use it even in restrictive environments.


At best you'd be able to prevent some part from working easily but since the plan is to be able to swap out components, you'd still be able to utilize it by getting around blocks. Things like cjdns and message passing will make it very difficult to prevent data access.


Version 0.4.12 has a release candidate out now. Here's part of the changelog:

>Ipfs 0.4.12 brings with it many important fixes for the huge spike in network size we've seen this past month. These changes include the Connection Manager, faster batching in ipfs add, libp2p fixes that reduce CPU usage, and a bunch of new documentation.

>The most critical change is the 'Connection Manager': it allows an ipfs node to maintain a limited set of connections to other peers in the network. By default (and with no config changes required by the user), ipfs nodes will now try to maintain between 600 and 900 open connections. These limits are still likely higher than needed, and future releases may lower the default recommendation, but for now we want to make changes gradually. The rationale for this selection of numbers is as follows:

>>The DHT routing table for a large network may rise to around 400 peers

>>Bitswap connections tend to be separate from the DHT

>>PubSub connections also generally are another distinct set of peers (including js-ipfs nodes)

>Because of this, we selected 600 as a 'LowWater' number, and 900 as a 'HighWater' number to avoid having to clear out connections too frequently. You can configure different numbers as you see fit via the Swarm.ConnMgr field in your ipfs config file. See here for more details.

>Disk utilization during ipfs add has been optimized for large files by doing batch writes in parallel. Previously, when adding a large file, users might have noticed that the add progressed by about 8MB at a time, with brief pauses in between. This was caused by quickly filling up the batch, then blocking while it was writing to disk. We now write to disk in the background while continuing to add the remainder of the file.

>Other changes in this release have noticeably reduced memory consumption and CPU usage. This was done by optimising some frequently called functions in libp2p that were expensive in terms of both CPU usage and memory allocations. We also lowered the yamux accept buffer sizes which were raised over a year ago to combat a separate bug that has since been fixed.

>And finally, thank you to everyone who filed bugs, tested out the release candidates, filed pull requests, and contributed in any other way to this release!



The final version of 0.4.12 was released yesterday. Now that it won't spam my network as much I might host some content. But the real question is, is my bandwith is better spent seeding the torrents or seeding them on IPFS?



>torrents or IPFS

Do you think it's likely you'd max out with both running? I tend to have my torrent and ipfs daemons running simultaneously, if I get bittorrent traffic it's usually just a burst for an hour or a very slow trickle to someone on the other side of the world. Same with IPFS, most of the time it's idle, and I have a lot of content shared on both + other networks.

Right now I have both running + soulseek.



>communism: the protocol



A global network is something I'm okay with my computer contributing to. It's robot communism at best. The costs and benefits of a decentralized and/or distributed system are more appealing than the faults associated with a centralized system, in my opinion.

The details of IPFS itself (while not new) seem especially nice, things like immutability, content addressing, and the inherent lack of trust associated with P2P systems which encourages better validation and security practices at a network level.

Things like pinning services (essentially CDNs), private networks, and real-time dynamic content via pubsub, allow for some opportunity for capitalists as well.

I think things like Freenet which force you to share the load are more communistic, with IPFS you're only ever sharing what you yourself want to share.


Wew, page 9. Anyway...

Charlie Manson Superstar (ogv)



File: c866b4ca831775f⋯.jpg (10.78 KB, 372x268, 93:67, noire_yamero.jpg)




What a quality argument. It's good to know that the best and brightest are still here.


go-ipfs 0.4.13 is out. Judging by how quick it was, you should probably download it immediately if you're on 0.4.12 already.

>Ipfs 0.4.13 is a patch release that fixes two high priority issues that were discovered in the 0.4.12 release.


>Fix periodic bitswap deadlock (ipfs/go-ipfs#4386)

<If a sessions context gets canceled during an isInterested request, bitswap (all of bitswap) can deadlock.

>>Fix badgerds crash on startup (ipfs/go-ipfs#4384)

<After changing the datastore to badgerds you can no longer start a daemon without a panic.


File: 748518956b7769c⋯.jpg (9.02 KB, 479x163, 479:163, LfZpouS_d.jpg)


Because you arent contributing.


File: a5e79f0306adaf5⋯.jpg (23.43 KB, 400x300, 4:3, Charlie Manson Superstar.jpg)


Cool intro graphic.

Friendly reminder to the thread

ipfs add -w "Charlie Manson Superstar.ogv"


ipfs files mkdir /tmp
ipfs files cp /ipfs/QmdSCNSSdpS3j6vudHR87FChHHus4jsgMKGMKLP4g86tRM "/tmp/Charlie Manson Superstar.ogv"
ipfs stat /tmp
ipfs files rm -r /tmp


You don't need to fetch the file to craft that hash in the latter either, and you can always get the base hash back from it if you need it bare.

ipfs ls /ipfs/QmcaQifUM8ixuERUXe4fXX89hCAhgKK8MGFPomSEcZgn2C

I wonder if this linkifies with the extension



>pushing a release to fix an experimental feature



oh hi i hrd thr ws cp?




Nothing very interesting there.


File: d1915fa375efa98⋯.jpg (99.15 KB, 1280x720, 16:9, animegataris.jpg)

Animegataris. Episodes 1 to 10.




Is your daemon running? I can't access this.

also prefixed for the browser addon /ipfs/QmbNZZpQRThNurXNPhcazA2Gw52FCC3SxChiC3T6GiMB3e


Saving the thread while we wait for a new version, so I might as well ask: What is the difference between pin and add? I'm still not really clear on that.


no offense guys but.. why care about this shit? I can't use it through tor or i2p so it's just the same as torrents: SEND ME A DMCA PLZ! I want 3 strikes and my net cut off!



Pin tells the GC not to reap the blocks until you unpin them.




Reminder that by default, adding something pins it, if you don't want that use

ipfs add --pin=false
and whatever you just added will be reaped on the next GC, if you don't want it to be reaped, pin it with
ipfs pin add *hash*


So when are we expecting for DHT to work in js-ipfs, so we can properly run it on websites?

At the moment, running a node on a webpage only allows you to retrieve content that the default gateways seem to have stored.



Would it be possible to just continuously reannounce to keep the files cached in the gateways? It's a shitty solution but it might work as a stopgap until js-ipfs is more complete.





You are the ultimate brainlet. This isn't just for pirating shit, it's also for censorship circumvention and freedum.

Where is my in-page IPFS without gateways? Once we have that it will be a new age of Internet prosperity. Smugboard can finally replace this crumbling shell of an imageboard.



>Where is my in-page IPFS without gateways?

Eventually. It seems js-ipfs doesn't yet have proper routing or something, so it can't connect to the DHT, and or connect to nodes it hasn't already connected to. This limits js-ipfs to any node on the bootstrap list.

I was thinking it'd be fine to just use go-ipfs and have people download a client (which runs the daemon in the background), but seems like everyone these days wants to do everything from within the browser. Thus, we wait for js-ipfs to reach a usable state.



>Thus, we wait for js-ipfs to reach a usable state.

Yes, don't wanna blow our load before making it perfectly transparent for normies. They should use those filecoin shekels to hire like 20 more devs though, is this not the most important development in Internet history since social media? Where are the buzzwords, the hype? Oh right, this will kill CDNs and file hosting sites (jewtube etc) among other things.



>They should use those filecoin shekels to hire like 20 more devs though

I agree. Maybe I'd see it if I really analyzed how many people are making pull requests, but it seems like things are moving the same speed they always did. Granted the big things to tackle moving forward are executive decisions about how to implement complicated shit like bitswap, but you can pay people to help with bug fixing while you approach conceptual solutions. Filecoin can't live without IPFS and vice versa. It's very much a yin-yang thing. I'd like to see some serious muscle put into shipping out a 1.0 before Filecoin even hits open beta. God forbid they come out of the gate with critical bugs or scaling issues.

Another nice thing to see would be hiring someone to help develop community projects. Imagine a guy who knows the software back to front helping out all the little guys in that IPFS Shipyard, especially things that could work as building blocks for bigger projects.


File: 28be3b99e2c56ec⋯.webm (11.92 MB, 800x450, 16:9, IPFS Alpha _ Why We Must ….webm)


Bumping a thread on a slow board, with 15 pages of threads, after only 4 hours between the last bump, is bad form. Please don't do that.


I didn't want to just complain about wasting posts while wasting a post myself, so I'll give my input on this.

I agree that this is extremely important but important things aren't always exciting. IPFS enables us to reliably host data on our own, with all kinds of redundancy, distribution, censorship resistance, and a nice promise of practical permanence(as long as someone has the data, anyone else can access it using the original hash, forever(no dead links)). All of these things are indeed important but it's not exactly exciting, for some of us it's almost frustrating since we sit here and say "it should have always been this way...", with that in mind I can understand why there isn't much noticeable "hype", it's mostly silent experimentation and adoption which seems fine to me. I don't think IPFS needs any kind evangelizing, especially not right now. I think anyone that comes across it can come to their own conclusion easily on whether it's worth using or not, it will grow naturally regardless of it's public image, like most good technologies. What comes to my mind is things like BitTorrent, it's huge today and it's not because of any kind of marketing bs and without any promise of people making money off of it. Also it's a long way from being finished anyway.

That all being said it's not like there aren't people writing about IPFS and getting excited, the project creator does more than enough talks at conventions, schools, etc. and in my opinion is doing a good job explaining what it is, does, and why you should care about that.

The old threads used to link this a lot


Since 2015 I've seen more and more publications about it so it's not like it's not happening, it's just not mainstream yet, and I don't think it should be until it's finished and all polished up. I think they have a good balance going on right now, it's too early to hype things up when they're unfinished and changing rapidly, and what's nice is it feels like the community understands this too. Imagine other projects that get popular too early, they get talked up a lot but don't actually meet their promises yet, people hear about it, they try it, and they get disappointed.



When https://github.com/ipfs/js-ipfs/pull/856 is merged. Which is taking fucking forever.



>it feels like the community understands this too

This is true, it's pretty clear to me everyone seems to know what's up regarding this. Nobody wants to push this mainstream until it's something we can stand behind.


Someone want to explain to me how the

ipfs p2p
subcommands work?

If I'm understanding correctly, you can basically listen in for TCP/UDP connections, and connect to them by resolving an IPFS Peer-ID, instead of a domain or IP address?

Looks pretty useful; has anyone actually made use of these features?



It's experimental, so documentation is light. This thread has an easy example demonstration: https://github.com/ipfs/go-ipfs/issues/3994

>To enable the command, you will need to run:

<ipfs config --json Experimental.Libp2pStreamMounting true

>And restart your daemon.

>Basic usage of ipfs p2p is as follows:

>First, open up a p2p listener on one ipfs node:

<p2p listener open p2p-test /ip4/

>This will create a libp2p protocol handler for a protocol named p2p-test and then open up a tcp socket locally on port 10101. Your application can now connect to that port to interact with the service.

>On the other ipfs node, connect to the listener on node A

<ipfsi 1 p2p stream dial $NODE_A_PEERID p2p-test /ip4/

>Now the ipfs nodes are connected to eachother. This command also created another tcp socket for the other end of your application to connect to. Once the other end connects to that socket, both sides of your application should be able to communicate, with their traffic being routed through libp2p.

>The easiest way to try this out is with netcat.

I'm not sure what he meant by "ipfsi 1" but I assume it's a typo, and you should just change that to "ipfs".



Seems more fleshed out for js-ipfs, for reasons that are only natural.




libp2p is a set of protocols and utilities, of which IPFS utilizes most of them. "ipfs p2p" is a subcommand for binding a normal application to your PeerID via ports, so it's possible to create a centralized website on top of IPFS, with decentralized addressing.



Yeah, nevermind, it's pretty much the same thing, it looks like.


Or at least, it's a part of it.


File: 235592ec8e55aa0⋯.png (104.32 KB, 1856x931, 1856:931, 34874493-0d306e5a-f791-11e….png)

Never realized this existed until now. It's almost exactly what I had in mind for shilling IPFS to normalfags, especially ones on Windows.


Anyone try it?


File: ce020cd9cf6c9c1⋯.gif (70.31 KB, 800x350, 16:7, licecap-trim.gif)


I've been using this one



File: 260226e4690708a⋯.png (15 KB, 827x562, 827:562, asdf.PNG)




They have to do that on the main gateway to avoid getting vanned for knowingly hosting illegal content. If you want to link it to other people who don't have ipfs, you can use someone elses gateway or better yet, use your own with your own blacklist.


I never even looked into how to setup a blacklist for it though.

What's the content of that hash?



Found it on /ipfs/QmU5XsVwvJfTcCwqkK1SmTqDmXWSQWaTa7ZcVLY2PDxNxG/ipfs_links.html

It's labeled as "Programming books".


File: 4823163c2b4b10d⋯.png (2.97 MB, 2200x3276, 550:819, 1512610818106.png)


I picked a random one from here that doesn't use their DMCA list.



Remember to use your local gateway regardless.


Don't read Zed's books either.



Sipser's book isn't that bad. What are you getting at anon?



I had high hopes after the first category, which is spot on.

If your school teaches Java, it's basically a "Java school", there is literally no lower circle of hell for a CS school, except teaching C# maybe.

Second category : Papadimitriou is a respected computer scientist, was this book that bad?

Third category : Sipser taught at MIT for 30 years.

Fourth category : comparing apples and oranges, I very much doubt that books on the left are used at the same level or for the same course as books on the left.

I will do my own list, cool idea.


File: f8cadcc3c7db9d1⋯.jpeg (844.28 KB, 2200x3276, 550:819, good_bad_schools.jpeg)

[Return][Go to top][Catalog][Nerve Center][Cancer][Post a Reply]
Delete Post [ ]
[ / / / / / / / / / / / / / ] [ dir / ausneets / had / htg / leftpol / radcorp / strek / sw / thestorm ]