GitHub - freewil/node-bitcoin: UNMAINTAINED - See: https ...

The All-in-One, Cheap, Energy-efficient Package to Store and Mint Peercoin

Dedicated OS for low-grade and energy efficient HW like RaspberryPi or Beaglebone Black to run Peercoin minter so that decentralization is guaranteed
[link]

PSA: Just running the BTC core client won't really count as an actual node: Forward port 8333! /r/Bitcoin

PSA: Just running the BTC core client won't really count as an actual node: Forward port 8333! /Bitcoin submitted by BitcoinAllBot to BitcoinAll [link] [comments]

Gridcoin 5.0.0.0-Mandatory "Fern" Release

https://github.com/gridcoin-community/Gridcoin-Research/releases/tag/5.0.0.0
Finally! After over ten months of development and testing, "Fern" has arrived! This is a whopper. 240 pull requests merged. Essentially a complete rewrite that was started with the scraper (the "neural net" rewrite) in "Denise" has now been completed. Practically the ENTIRE Gridcoin specific codebase resting on top of the vanilla Bitcoin/Peercoin/Blackcoin vanilla PoS code has been rewritten. This removes the team requirement at last (see below), although there are many other important improvements besides that.
Fern was a monumental undertaking. We had to encode all of the old rules active for the v10 block protocol in new code and ensure that the new code was 100% compatible. This had to be done in such a way as to clear out all of the old spaghetti and ring-fence it with tightly controlled class implementations. We then wrote an entirely new, simplified ruleset for research rewards and reengineered contracts (which includes beacon management, polls, and voting) using properly classed code. The fundamentals of Gridcoin with this release are now on a very sound and maintainable footing, and the developers believe the codebase as updated here will serve as the fundamental basis for Gridcoin's future roadmap.
We have been testing this for MONTHS on testnet in various stages. The v10 (legacy) compatibility code has been running on testnet continuously as it was developed to ensure compatibility with existing nodes. During the last few months, we have done two private testnet forks and then the full public testnet testing for v11 code (the new protocol which is what Fern implements). The developers have also been running non-staking "sentinel" nodes on mainnet with this code to verify that the consensus rules are problem-free for the legacy compatibility code on the broader mainnet. We believe this amount of testing is going to result in a smooth rollout.
Given the amount of changes in Fern, I am presenting TWO changelogs below. One is high level, which summarizes the most significant changes in the protocol. The second changelog is the detailed one in the usual format, and gives you an inkling of the size of this release.

Highlights

Protocol

Note that the protocol changes will not become active until we cross the hard-fork transition height to v11, which has been set at 2053000. Given current average block spacing, this should happen around October 4, about one month from now.
Note that to get all of the beacons in the network on the new protocol, we are requiring ALL beacons to be validated. A two week (14 day) grace period is provided by the code, starting at the time of the transition height, for people currently holding a beacon to validate the beacon and prevent it from expiring. That means that EVERY CRUNCHER must advertise and validate their beacon AFTER the v11 transition (around Oct 4th) and BEFORE October 18th (or more precisely, 14 days from the actual date of the v11 transition). If you do not advertise and validate your beacon by this time, your beacon will expire and you will stop earning research rewards until you advertise and validate a new beacon. This process has been made much easier by a brand new beacon "wizard" that helps manage beacon advertisements and renewals. Once a beacon has been validated and is a v11 protocol beacon, the normal 180 day expiration rules apply. Note, however, that the 180 day expiration on research rewards has been removed with the Fern update. This means that while your beacon might expire after 180 days, your earned research rewards will be retained and can be claimed by advertising a beacon with the same CPID and going through the validation process again. In other words, you do not lose any earned research rewards if you do not stake a block within 180 days and keep your beacon up-to-date.
The transition height is also when the team requirement will be relaxed for the network.

GUI

Besides the beacon wizard, there are a number of improvements to the GUI, including new UI transaction types (and icons) for staking the superblock, sidestake sends, beacon advertisement, voting, poll creation, and transactions with a message. The main screen has been revamped with a better summary section, and better status icons. Several changes under the hood have improved GUI performance. And finally, the diagnostics have been revamped.

Blockchain

The wallet sync speed has been DRASTICALLY improved. A decent machine with a good network connection should be able to sync the entire mainnet blockchain in less than 4 hours. A fast machine with a really fast network connection and a good SSD can do it in about 2.5 hours. One of our goals was to reduce or eliminate the reliance on snapshots for mainnet, and I think we have accomplished that goal with the new sync speed. We have also streamlined the in-memory structures for the blockchain which shaves some memory use.
There are so many goodies here it is hard to summarize them all.
I would like to thank all of the contributors to this release, but especially thank @cyrossignol, whose incredible contributions formed the backbone of this release. I would also like to pay special thanks to @barton2526, @caraka, and @Quezacoatl1, who tirelessly helped during the testing and polishing phase on testnet with testing and repeated builds for all architectures.
The developers are proud to present this release to the community and we believe this represents the starting point for a true renaissance for Gridcoin!

Summary Changelog

Accrual

Changed

Most significantly, nodes calculate research rewards directly from the magnitudes in EACH superblock between stakes instead of using a two- or three- point average based on a CPID's current magnitude and the magnitude for the CPID when it last staked. For those long-timers in the community, this has been referred to as "Superblock Windows," and was first done in proof-of-concept form by @denravonska.

Removed

Beacons

Added

Changed

Removed

Unaltered

As a reminder:

Superblocks

Added

Changed

Removed

Voting

Added

Changed

Removed

Detailed Changelog

[5.0.0.0] 2020-09-03, mandatory, "Fern"

Added

Changed

Removed

Fixed

submitted by jamescowens to gridcoin [link] [comments]

Scaling Reddit Community Points with Arbitrum Rollup: a piece of cake

Scaling Reddit Community Points with Arbitrum Rollup: a piece of cake
https://preview.redd.it/b80c05tnb9e51.jpg?width=2550&format=pjpg&auto=webp&s=850282c1a3962466ed44f73886dae1c8872d0f31
Submitted for consideration to The Great Reddit Scaling Bake-Off
Baked by the pastry chefs at Offchain Labs
Please send questions or comments to [[email protected] ](mailto:[email protected])
1. Overview
We're excited to submit Arbitrum Rollup for consideration to The Great Reddit Scaling Bake-Off. Arbitrum Rollup is the only Ethereum scaling solution that supports arbitrary smart contracts without compromising on Ethereum's security or adding points of centralization. For Reddit, this means that Arbitrum can not only scale the minting and transfer of Community Points, but it can foster a creative ecosystem built around Reddit Community Points enabling points to be used in a wide variety of third party applications. That's right -- you can have your cake and eat it too!
Arbitrum Rollup isn't just Ethereum-style. Its Layer 2 transactions are byte-for-byte identical to Ethereum, which means Ethereum users can continue to use their existing addresses and wallets, and Ethereum developers can continue to use their favorite toolchains and development environments out-of-the-box with Arbitrum. Coupling Arbitrum’s tooling-compatibility with its trustless asset interoperability, Reddit not only can scale but can onboard the entire Ethereum community at no cost by giving them the same experience they already know and love (well, certainly know).
To benchmark how Arbitrum can scale Reddit Community Points, we launched the Reddit contracts on an Arbitrum Rollup chain. Since Arbitrum provides full Solidity support, we didn't have to rewrite the Reddit contracts or try to mimic their functionality using an unfamiliar paradigm. Nope, none of that. We launched the Reddit contracts unmodified on Arbitrum Rollup complete with support for minting and distributing points. Like every Arbitrum Rollup chain, the chain included a bridge interface in which users can transfer Community Points or any other asset between the L1 and L2 chains. Arbitrum Rollup chains also support dynamic contract loading, which would allow third-party developers to launch custom ecosystem apps that integrate with Community Points on the very same chain that runs the Reddit contracts.
1.1 Why Ethereum
Perhaps the most exciting benefit of distributing Community Points using a blockchain is the ability to seamlessly port points to other applications and use them in a wide variety of contexts. Applications may include simple transfers such as a restaurant that allows Redditors to spend points on drinks. Or it may include complex smart contracts -- such as placing Community Points as a wager for a multiparty game or as collateral in a financial contract.
The common denominator between all of the fun uses of Reddit points is that it needs a thriving ecosystem of both users and developers, and the Ethereum blockchain is perhaps the only smart contract platform with significant adoption today. While many Layer 1 blockchains boast lower cost or higher throughput than the Ethereum blockchain, more often than not, these attributes mask the reality of little usage, weaker security, or both.
Perhaps another platform with significant usage will rise in the future. But today, Ethereum captures the mindshare of the blockchain community, and for Community Points to provide the most utility, the Ethereum blockchain is the natural choice.
1.2 Why Arbitrum
While Ethereum's ecosystem is unmatched, the reality is that fees are high and capacity is too low to support the scale of Reddit Community Points. Enter Arbitrum. Arbitrum Rollup provides all of the ecosystem benefits of Ethereum, but with orders of magnitude more capacity and at a fraction of the cost of native Ethereum smart contracts. And most of all, we don't change the experience from users. They continue to use the same wallets, addresses, languages, and tools.
Arbitrum Rollup is not the only solution that can scale payments, but it is the only developed solution that can scale both payments and arbitrary smart contracts trustlessly, which means that third party users can build highly scalable add-on apps that can be used without withdrawing money from the Rollup chain. If you believe that Reddit users will want to use their Community Points in smart contracts--and we believe they will--then it makes the most sense to choose a single scaling solution that can support the entire ecosystem, eliminating friction for users.
We view being able to run smart contracts in the same scaling solution as fundamentally critical since if there's significant demand in running smart contracts from Reddit's ecosystem, this would be a load on Ethereum and would itself require a scaling solution. Moreover, having different scaling solutions for the minting/distribution/spending of points and for third party apps would be burdensome for users as they'd have to constantly shuffle their Points back and forth.
2. Arbitrum at a glance
Arbitrum Rollup has a unique value proposition as it offers a combination of features that no other scaling solution achieves. Here we highlight its core attributes.
Decentralized. Arbitrum Rollup is as decentralized as Ethereum. Unlike some other Layer 2 scaling projects, Arbitrum Rollup doesn't have any centralized components or centralized operators who can censor users or delay transactions. Even in non-custodial systems, centralized components provide a risk as the operators are generally incentivized to increase their profit by extracting rent from users often in ways that severely degrade user experience. Even if centralized operators are altruistic, centralized components are subject to hacking, coercion, and potential liability.
Massive Scaling. Arbitrum achieves order of magnitude scaling over Ethereum's L1 smart contracts. Our software currently supports 453 transactions-per-second for basic transactions (at 1616 Ethereum gas per tx). We have a lot of room left to optimize (e.g. aggregating signatures), and over the next several months capacity will increase significantly. As described in detail below, Arbitrum can easily support and surpass Reddit's anticipated initial load, and its capacity will continue to improve as Reddit's capacity needs grow.
Low cost. The cost of running Arbitrum Rollup is quite low compared to L1 Ethereum and other scaling solutions such as those based on zero-knowledge proofs. Layer 2 fees are low, fixed, and predictable and should not be overly burdensome for Reddit to cover. Nobody needs to use special equipment or high-end machines. Arbitrum requires validators, which is a permissionless role that can be run on any reasonable on-line machine. Although anybody can act as a validator, in order to protect against a “tragedy of the commons” and make sure reputable validators are participating, we support a notion of “invited validators” that are compensated for their costs. In general, users pay (low) fees to cover the invited validators’ costs, but we imagine that Reddit may cover this cost for its users. See more on the costs and validator options below.
Ethereum Developer Experience. Not only does Arbitrum support EVM smart contracts, but the developer experience is identical to that of L1 Ethereum contracts and fully compatible with Ethereum tooling. Developers can port existing Solidity apps or write new ones using their favorite and familiar toolchains (e.g. Truffle, Buidler). There are no new languages or coding paradigms to learn.
Ethereum wallet compatibility. Just as in Ethereum, Arbitrum users need only hold keys, but do not have to store any coin history or additional data to protect or access their funds. Since Arbitrum transactions are semantically identical to Ethereum L1 transactions, existing Ethereum users can use their existing Ethereum keys with their existing wallet software such as Metamask.
Token interoperability. Users can easily transfer their ETH, ERC-20 and ERC-721 tokens between Ethereum and the Arbitrum Rollup chain. As we explain in detail below, it is possible to mint tokens in L2 that can subsequently be withdrawn and recognized by the L1 token contract.
Fast finality. Transactions complete with the same finality time as Ethereum L1 (and it's possible to get faster finality guarantees by trading away trust assumptions; see the Arbitrum Rollup whitepaper for details).
Non-custodial. Arbitrum Rollup is a non-custodial scaling solution, so users control their funds/points and neither Reddit nor anyone else can ever access or revoke points held by users.
Censorship Resistant. Since it's completely decentralized, and the Arbitrum protocol guarantees progress trustlessly, Arbitrum Rollup is just as censorship-proof as Ethereum.
Block explorer. The Arbitrum Rollup block explorer allows users to view and analyze transactions on the Rollup chain.
Limitations
Although this is a bake-off, we're not going to sugar coat anything. Arbitrum Rollup, like any Optimistic Rollup protocol, does have one limitation, and that's the delay on withdrawals.
As for the concrete length of the delay, we've done a good deal of internal modeling and have blogged about this as well. Our current modeling suggests a 3-hour delay is sufficient (but as discussed in the linked post there is a tradeoff space between the length of the challenge period and the size of the validators’ deposit).
Note that this doesn't mean that the chain is delayed for three hours. Arbitrum Rollup supports pipelining of execution, which means that validators can keep building new states even while previous ones are “in the pipeline” for confirmation. As the challenge delays expire for each update, a new state will be confirmed (read more about this here).
So activity and progress on the chain are not delayed by the challenge period. The only thing that's delayed is the consummation of withdrawals. Recall though that any single honest validator knows immediately (at the speed of L1 finality) which state updates are correct and can guarantee that they will eventually be confirmed, so once a valid withdrawal has been requested on-chain, every honest party knows that the withdrawal will definitely happen. There's a natural place here for a liquidity market in which a validator (or someone who trusts a validator) can provide withdrawal loans for a small interest fee. This is a no-risk business for them as they know which withdrawals will be confirmed (and can force their confirmation trustlessly no matter what anyone else does) but are just waiting for on-chain finality.
3. The recipe: How Arbitrum Rollup works
For a description of the technical components of Arbitrum Rollup and how they interact to create a highly scalable protocol with a developer experience that is identical to Ethereum, please refer to the following documents:
Arbitrum Rollup Whitepaper
Arbitrum academic paper (describes a previous version of Arbitrum)
4. Developer docs and APIs
For full details about how to set up and interact with an Arbitrum Rollup chain or validator, please refer to our developer docs, which can be found at https://developer.offchainlabs.com/.
Note that the Arbitrum version described on that site is older and will soon be replaced by the version we are entering in Reddit Bake-Off, which is still undergoing internal testing before public release.
5. Who are the validators?
As with any Layer 2 protocol, advancing the protocol correctly requires at least one validator (sometimes called block producers) that is honest and available. A natural question is: who are the validators?
Recall that the validator set for an Arbitrum chain is open and permissionless; anyone can start or stop validating at will. (A useful analogy is to full nodes on an L1 chain.) But we understand that even though anyone can participate, Reddit may want to guarantee that highly reputable nodes are validating their chain. Reddit may choose to validate the chain themselves and/or hire third-party validators.To this end, we have begun building a marketplace for validator-for-hire services so that dapp developers can outsource validation services to reputable nodes with high up-time. We've announced a partnership in which Chainlink nodes will provide Arbitrum validation services, and we expect to announce more partnerships shortly with other blockchain infrastructure providers.
Although there is no requirement that validators are paid, Arbitrum’s economic model tracks validators’ costs (e.g. amount of computation and storage) and can charge small fees on user transactions, using a gas-type system, to cover those costs. Alternatively, a single party such as Reddit can agree to cover the costs of invited validators.
6. Reddit Contract Support
Since Arbitrum contracts and transactions are byte-for-byte compatible with Ethereum, supporting the Reddit contracts is as simple as launching them on an Arbitrum chain.
Minting. Arbitrum Rollup supports hybrid L1/L2 tokens which can be minted in L2 and then withdrawn onto the L1. An L1 contract at address A can make a special call to the EthBridge which deploys a "buddy contract" to the same address A on an Arbitrum chain. Since it's deployed at the same address, users can know that the L2 contract is the authorized "buddy" of the L1 contract on the Arbitrum chain.
For minting, the L1 contract is a standard ERC-20 contract which mints and burns tokens when requested by the L2 contract. It is paired with an ERC-20 contract in L2 which mints tokens based on whatever programmer provided minting facility is desired and burns tokens when they are withdrawn from the rollup chain. Given this base infrastructure, Arbitrum can support any smart contract based method for minting tokens in L2, and indeed we directly support Reddit's signature/claim based minting in L2.
Batch minting. What's better than a mint cookie? A whole batch! In addition to supporting Reddit’s current minting/claiming scheme, we built a second minting design, which we believe outperforms the signature/claim system in many scenarios.
In the current system, Reddit periodically issues signed statements to users, who then take those statements to the blockchain to claim their tokens. An alternative approach would have Reddit directly submit the list of users/amounts to the blockchain and distribute the tokens to the users without the signature/claim process.
To optimize the cost efficiency of this approach, we designed an application-specific compression scheme to minimize the size of the batch distribution list. We analyzed the data from Reddit's previous distributions and found that the data is highly compressible since token amounts are small and repeated, and addresses appear multiple times. Our function groups transactions by size, and replaces previously-seen addresses with a shorter index value. We wrote client code to compress the data, wrote a Solidity decompressing function, and integrated that function into Reddit’s contract running on Arbitrum.
When we ran the compression function on the previous Reddit distribution data, we found that we could compress batched minting data down to to 11.8 bytes per minting event (averaged over a 6-month trace of Reddit’s historical token grants)compared with roughly 174 bytes of on-chain data needed for the signature claim approach to minting (roughly 43 for an RLP-encoded null transaction + 65 for Reddit's signature + 65 for the user's signature + roughly 8 for the number of Points) .
The relative benefit of the two approaches with respect to on-chain call data cost depends on the percentage of users that will actually claim their tokens on chain. With the above figures, batch minting will be cheaper if roughly 5% of users redeem their claims. We stress that our compression scheme is not Arbitrum-specific and would be beneficial in any general-purpose smart contract platform.
8. Benchmarks and costs
In this section, we give the full costs of operating the Reddit contracts on an Arbitrum Rollup chain including the L1 gas costs for the Rollup chain, the costs of computation and storage for the L2 validators as well as the capital lockup requirements for staking.
Arbitrum Rollup is still on testnet, so we did not run mainnet benchmarks. Instead, we measured the L1 gas cost and L2 workload for Reddit operations on Arbitrum and calculated the total cost assuming current Ethereum gas prices. As noted below in detail, our measurements do not assume that Arbitrum is consuming the entire capacity of Ethereum. We will present the details of our model now, but for full transparency you can also play around with it yourself and adjust the parameters, by copying the spreadsheet found here.
Our cost model is based on measurements of Reddit’s contracts, running unmodified (except for the addition of a batch minting function) on Arbitrum Rollup on top of Ethereum.
On the distribution of transactions and frequency of assertions. Reddit's instructions specify the following minimum parameters that submissions should support:
Over a 5 day period, your scaling PoC should be able to handle:
  • 100,000 point claims (minting & distributing points)
  • 25,000 subscriptions
  • 75,000 one-off points burning
  • 100,000 transfers
We provide the full costs of operating an Arbitrum Rollup chain with this usage under the assumption that tokens are minted or granted to users in batches, but other transactions are uniformly distributed over the 5 day period. Unlike some other submissions, we do not make unrealistic assumptions that all operations can be submitted in enormous batches. We assume that batch minting is done in batches that use only a few percent on an L1 block’s gas, and that other operations come in evenly over time and are submitted in batches, with one batch every five minutes to keep latency reasonable. (Users are probably already waiting for L1 finality, which takes at least that long to achieve.)
We note that assuming that there are only 300,000 transactions that arrive uniformly over the 5 day period will make our benchmark numbers lower, but we believe that this will reflect the true cost of running the system. To see why, say that batches are submitted every five minutes (20 L1 blocks) and there's a fixed overhead of c bytes of calldata per batch, the cost of which will get amortized over all transactions executed in that batch. Assume that each individual transaction adds a marginal cost of t. Lastly assume the capacity of the scaling system is high enough that it can support all of Reddit's 300,000 transactions within a single 20-block batch (i.e. that there is more than c + 300,000*t byes of calldata available in 20 blocks).
Consider what happens if c, the per-batch overhead, is large (which it is in some systems, but not in Arbitrum). In the scenario that transactions actually arrive at the system's capacity and each batch is full, then c gets amortized over 300,000 transactions. But if we assume that the system is not running at capacity--and only receives 300,000 transactions arriving uniformly over 5 days-- then each 20-block assertion will contain about 200 transactions, and thus each transaction will pay a nontrivial cost due to c.
We are aware that other proposals presented scaling numbers assuming that 300,000 transactions arrived at maximum capacity and was executed in a single mega-transaction, but according to our estimates, for at least one such report, this led to a reported gas price that was 2-3 orders of magnitude lower than it would have been assuming uniform arrival. We make more realistic batching assumptions, and we believe Arbitrum compares well when batch sizes are realistic.
Our model. Our cost model includes several sources of cost:
  • L1 gas costs: This is the cost of posting transactions as calldata on the L1 chain, as well as the overhead associated with each batch of transactions, and the L1 cost of settling transactions in the Arbitrum protocol.
  • Validator’s staking costs: In normal operation, one validator will need to be staked. The stake is assumed to be 0.2% of the total value of the chain (which is assumed to be $1 per user who is eligible to claim points). The cost of staking is the interest that could be earned on the money if it were not staked.
  • Validator computation and storage: Every validator must do computation to track the chain’s processing of transactions, and must maintain storage to keep track of the contracts’ EVM storage. The cost of computation and storage are estimated based on measurements, with the dollar cost of resources based on Amazon Web Services pricing.
It’s clear from our modeling that the predominant cost is for L1 calldata. This will probably be true for any plausible rollup-based system.
Our model also shows that Arbitrum can scale to workloads much larger than Reddit’s nominal workload, without exhausting L1 or L2 resources. The scaling bottleneck will ultimately be calldata on the L1 chain. We believe that cost could be reduced substantially if necessary by clever encoding of data. (In our design any compression / decompression of L2 transaction calldata would be done by client software and L2 programs, never by an L1 contract.)
9. Status of Arbitrum Rollup
Arbitrum Rollup is live on Ethereum testnet. All of the code written to date including everything included in the Reddit demo is open source and permissively licensed under the Apache V2 license. The first testnet version of Arbitrum Rollup was released on testnet in February. Our current internal version, which we used to benchmark the Reddit contracts, will be released soon and will be a major upgrade.
Both the Arbitrum design as well as the implementation are heavily audited by independent third parties. The Arbitrum academic paper was published at USENIX Security, a top-tier peer-reviewed academic venue. For the Arbitrum software, we have engaged Trail of Bits for a security audit, which is currently ongoing, and we are committed to have a clean report before launching on Ethereum mainnet.
10. Reddit Universe Arbitrum Rollup Chain
The benchmarks described in this document were all measured using the latest internal build of our software. When we release the new software upgrade publicly we will launch a Reddit Universe Arbitrum Rollup chain as a public demo, which will contain the Reddit contracts as well as a Uniswap instance and a Connext Hub, demonstrating how Community Points can be integrated into third party apps. We will also allow members of the public to dynamically launch ecosystem contracts. We at Offchain Labs will cover the validating costs for the Reddit Universe public demo.
If the folks at Reddit would like to evaluate our software prior to our public demo, please email us at [email protected] and we'd be more than happy to provide early access.
11. Even more scaling: Arbitrum Sidechains
Rollups are an excellent approach to scaling, and we are excited about Arbitrum Rollup which far surpasses Reddit's scaling needs. But looking forward to Reddit's eventual goal of supporting hundreds of millions of users, there will likely come a time when Reddit needs more scaling than any Rollup protocol can provide.
While Rollups greatly reduce costs, they don't break the linear barrier. That is, all transactions have an on-chain footprint (because all calldata must be posted on-chain), albeit a far smaller one than on native Ethereum, and the L1 limitations end up being the bottleneck for capacity and cost. Since Ethereum has limited capacity, this linear use of on-chain resources means that costs will eventually increase superlinearly with traffic.
The good news is that we at Offchain Labs have a solution in our roadmap that can satisfy this extreme-scaling setting as well: Arbitrum AnyTrust Sidechains. Arbitrum Sidechains are similar to Arbitrum Rollup, but deviate in that they name a permissioned set of validators. When a chain’s validators agree off-chain, they can greatly reduce the on-chain footprint of the protocol and require almost no data to be put on-chain. When validators can't reach unanimous agreement off-chain, the protocol reverts to Arbitrum Rollup. Technically, Arbitrum Sidechains can be viewed as a hybrid between state channels and Rollup, switching back and forth as necessary, and combining the performance and cost that state channels can achieve in the optimistic case, with the robustness of Rollup in other cases. The core technical challenge is how to switch seamlessly between modes and how to guarantee that security is maintained throughout.
Arbitrum Sidechains break through this linear barrier, while still maintaining a high level of security and decentralization. Arbitrum Sidechains provide the AnyTrust guarantee, which says that as long as any one validator is honest and available (even if you don't know which one will be), the L2 chain is guaranteed to execute correctly according to its code and guaranteed to make progress. Unlike in a state channel, offchain progress does not require unanimous consent, and liveness is preserved as long as there is a single honest validator.
Note that the trust model for Arbitrum Sidechains is much stronger than for typical BFT-style chains which introduce a consensus "voting" protocols among a small permissioned group of validators. BFT-based protocols require a supermajority (more than 2/3) of validators to agree. In Arbitrum Sidechains, by contrast, all you need is a single honest validator to achieve guaranteed correctness and progress. Notice that in Arbitrum adding validators strictly increases security since the AnyTrust guarantee provides correctness as long as any one validator is honest and available. By contrast, in BFT-style protocols, adding nodes can be dangerous as a coalition of dishonest nodes can break the protocol.
Like Arbitrum Rollup, the developer and user experiences for Arbitrum Sidechains will be identical to that of Ethereum. Reddit would be able to choose a large and diverse set of validators, and all that they would need to guarantee to break through the scaling barrier is that a single one of them will remain honest.
We hope to have Arbitrum Sidechains in production in early 2021, and thus when Reddit reaches the scale that surpasses the capacity of Rollups, Arbitrum Sidechains will be waiting and ready to help.
While the idea to switch between channels and Rollup to get the best of both worlds is conceptually simple, getting the details right and making sure that the switch does not introduce any attack vectors is highly non-trivial and has been the subject of years of our research (indeed, we were working on this design for years before the term Rollup was even coined).
12. How Arbitrum compares
We include a comparison to several other categories as well as specific projects when appropriate. and explain why we believe that Arbitrum is best suited for Reddit's purposes. We focus our attention on other Ethereum projects.
Payment only Rollups. Compared to Arbitrum Rollup, ZK-Rollups and other Rollups that only support token transfers have several disadvantages:
  • As outlined throughout the proposal, we believe that the entire draw of Ethereum is in its rich smart contracts support which is simply not achievable with today's zero-knowledge proof technology. Indeed, scaling with a ZK-Rollup will add friction to the deployment of smart contracts that interact with Community Points as users will have to withdraw their coins from the ZK-Rollup and transfer them to a smart contract system (like Arbitrum). The community will be best served if Reddit builds on a platform that has built-in, frictionless smart-contract support.
  • All other Rollup protocols of which we are aware employ a centralized operator. While it's true that users retain custody of their coins, the centralized operator can often profit from censoring, reordering, or delaying transactions. A common misconception is that since they're non-custodial protocols, a centralized sequencer does not pose a risk but this is incorrect as the sequencer can wreak havoc or shake down users for side payments without directly stealing funds.
  • Sidechain type protocols can eliminate some of these issues, but they are not trustless. Instead, they require trust in some quorum of a committee, often requiring two-third of the committee to be honest, compared to rollup protocols like Arbitrum that require only a single honest party. In addition, not all sidechain type protocols have committees that are diverse, or even non-centralized, in practice.
  • Plasma-style protocols have a centralized operator and do not support general smart contracts.
13. Concluding Remarks
While it's ultimately up to the judges’ palate, we believe that Arbitrum Rollup is the bakeoff choice that Reddit kneads. We far surpass Reddit's specified workload requirement at present, have much room to optimize Arbitrum Rollup in the near term, and have a clear path to get Reddit to hundreds of millions of users. Furthermore, we are the only project that gives developers and users the identical interface as the Ethereum blockchain and is fully interoperable and tooling-compatible, and we do this all without any new trust assumptions or centralized components.
But no matter how the cookie crumbles, we're glad to have participated in this bake-off and we thank you for your consideration.
About Offchain Labs
Offchain Labs, Inc. is a venture-funded New York company that spun out of Princeton University research, and is building the Arbitrum platform to usher in the next generation of scalable, interoperable, and compatible smart contracts. Offchain Labs is backed by Pantera Capital, Compound VC, Coinbase Ventures, and others.
Leadership Team
Ed Felten
Ed Felten is Co-founder and Chief Scientist at Offchain Labs. He is on leave from Princeton University, where he is the Robert E. Kahn Professor of Computer Science and Public Affairs. From 2015 to 2017 he served at the White House as Deputy United States Chief Technology Officer and senior advisor to the President. He is an ACM Fellow and member of the National Academy of Engineering. Outside of work, he is an avid runner, cook, and L.A. Dodgers fan.
Steven Goldfeder
Steven Goldfeder is Co-founder and Chief Executive Officer at Offchain Labs. He holds a PhD from Princeton University, where he worked at the intersection of cryptography and cryptocurrencies including threshold cryptography, zero-knowledge proof systems, and post-quantum signatures. He is a co-author of Bitcoin and Cryptocurrency Technologies, the leading textbook on cryptocurrencies, and he has previously worked at Google and Microsoft Research, where he co-invented the Picnic signature algorithm. When not working, you can find Steven spending time with his family, taking a nature walk, or twisting balloons.
Harry Kalodner
Harry Kalodner is Co-founder and Chief Technology Officer at Offchain Labs where he leads the engineering team. Before the company he attended Princeton as a Ph.D candidate where his research explored economics, anonymity, and incentive compatibility of cryptocurrencies, and he also has worked at Apple. When not up at 3:00am writing code, Harry occasionally sleeps.
submitted by hkalodner to ethereum [link] [comments]

Fun with Dynamic DNS services and bitcoind

I realize that bitcoind has this capability built in, but thought it might be fun to configure it manually. As some background, most people look at their network and they have some address like 192.168.1.105. That is a private network address. Sites like WhatIsMyIpAddress.com will tell you what your EXTERNAL IP address is. When your running a bitcoin node, people will connect to your external IP.
Problem is... your ISP may cycle that IP address every week or so. bitcoind has a nice feature baked in called discover which will do the work of guessing your external IP as it changes. But if you are doing anything where you need to expose your RPC API externally (*danger*) then it would be nice to have some FQDN to go after to do the IP translation.
Originally, I had always solved this by exposing my node as an onion node. Since onion nodes traverse NATs and firewalls fairly effortlessly this was an easy option. But previously I had played around with Dynamic DNS service. I've used noip.com but there are likely dozens of services out there that will get the job done.
Once you have enabled a NoIP hostname, you can name it in your bitcoin.conf using the externalip=coolbtcnode.ddns.net. Then, assuming you setup all the RPC auth and port forwarding, you could access your node via RPC at coolbtcnode.ddns.net
As I said... it's a redundancy, since discover=1 does most of this for you, but still thought someone might enjoy the tip.
References:
submitted by brianddk to Bitcoin [link] [comments]

Thanks to all who submitted questions for Shiv Malik in the GAINS AMA yesterday, it was great to see so much interest in Data Unions! You can read the full transcript here:

Thanks to all who submitted questions for Shiv Malik in the GAINS AMA yesterday, it was great to see so much interest in Data Unions! You can read the full transcript here:

Gains x Streamr AMA Recap

https://preview.redd.it/o74jlxia8im51.png?width=1236&format=png&auto=webp&s=93eb37a3c9ed31dc3bf31c91295c6ee32e1582be
Thanks to everyone in our community who attended the GAINS AMA yesterday with, Shiv Malik. We were excited to see that so many people attended and gladly overwhelmed by the amount of questions we got from you on Twitter and Telegram. We decided to do a little recap of the session for anyone who missed it, and to archive some points we haven’t previously discussed with our community. Happy reading and thanks to Alexandre and Henry for having us on their channel!
What is the project about in a few simple sentences?
At Streamr we are building a real-time network for tomorrow’s data economy. It’s a decentralized, peer-to-peer network which we are hoping will one day replace centralized message brokers like Amazon’s AWS services. On top of that one of the things I’m most excited about are Data Unions. With Data Unions anyone can join the data economy and start monetizing the data they already produce. Streamr’s Data Union framework provides a really easy way for devs to start building their own data unions and can also be easily integrated into any existing apps.
Okay, sounds interesting. Do you have a concrete example you could give us to make it easier to understand?
The best example of a Data Union is the first one that has been built out of our stack. It's called Swash and it's a browser plugin.
You can download it here: http://swashapp.io/
And basically it helps you monetize the data you already generate (day in day out) as you browse the web. It's the sort of data that Google already knows about you. But this way, with Swash, you can actually monetize it yourself. The more people that join the union, the more powerful it becomes and the greater the rewards are for everyone as the data product sells to potential buyers.
Very interesting. What stage is the project/product at? It's live, right?
Yes. It's live. And the Data Union framework is in public beta. The Network is on course to be fully decentralized at some point next year.
How much can a regular person browsing the Internet expect to make for example?
So that's a great question. The answer is no one quite knows yet. We do know that this sort of data (consumer insights) is worth hundreds of millions and really isn't available in high quality. So With a union of a few million people, everyone could be getting 20-50 dollars a year. But it'll take a few years at least to realise that growth. Of course Swash is just one data union amongst many possible others (which are now starting to get built out on our platform!)
With Swash, I believe they now have 3,000 members. They need to get to 50,000 before they become really viable but they are yet to do any marketing. So all that is organic growth.
I assume the data is anonymized btw?
Yes. And there in fact a few privacy protecting tools Swash supplys to its users.
How does Swash compare to Brave?
So Brave really is about consent for people's attention and getting paid for that. They don't sell your data as such.
Swash can of course be a plugin with Brave and therefore you can make passive income browsing the internet. Whilst also then consenting to advertising if you so want to earn BAT.
Of course it's Streamr that is powering Swash. And we're looking at powering other DUs - say for example mobile applications.
The holy grail might be having already existing apps and platforms out there, integrating DU tech into their apps so people can consent (or not) to having their data sold - and then getting a cut of that revenue when it does sell.
The other thing to recognise is that the big tech companies monopolise data on a vast scale - data that we of course produce for them. That is stifling innovation.
Take for example a competitor map app. To effectively compete with Google maps or Waze, they need millions of users feeding real time data into it.
Without that - it's like Google maps used to be - static and a bit useless.
Right, so how do you convince these big tech companies that are producing these big apps to integrate with Streamr? Does it mean they wouldn't be able to monetize data as well on their end if it becomes more available through an aggregation of individuals?
If a map application does manage to scale to that level then inevitably Google buys them out - that's what happened with Waze.
But if you have a data union which bundles together the raw location data of millions of people then any application builder can come along and license that data for their app. This encourages all sorts of innovation and breaks the monopoly.
We're currently having conversations with Mobile Network operators to see if they want to pilot this new approach to data monetization. And that's what even more exciting. Just be explicit with users - do you want to sell your data? Okay, if yes, then which data point do you want to sell.
Then the mobile network operator (like T-mobile for example) then organises the sale of the data of those who consent and everyone gets a cut.
Streamr - in this example provides the backend to port and bundle the data, and also the token and payment rail for the payments.
So for big companies (mobile operators in this case), it's less logistics, handing over the implementation to you, and simply taking a cut?
It's a vision that we'll be able to talk more about more concretely in a few weeks time 😁
Compared to having to make sense of that data themselves (in the past) and selling it themselves
Sort of.
We provide the backened to port the data and the template smart contracts to distribute the payments.
They get to focus on finding buyers for the data and ensuring that the data that is being collected from the app is the kind of data that is valuable and useful to the world.
(Through our sister company TX, we also help build out the applications for them and ensure a smooth integration).
The other thing to add is that the reason why this vision is working, is that the current data economy is under attack. Not just from privacy laws such as GDPR, but also from Google shutting down cookies, bidstream data being investigated by the FTC (for example) and Apple making changes to IoS14 to make third party data sharing more explicit for users.
All this means that the only real places for thousands of multinationals to buy the sort of consumer insights they need to ensure good business decisions will be owned by Google/FB etc, or from SDKs or through this method - from overt, rich, consent from the consumer in return for a cut of the earnings.
A couple of questions to get a better feel about Streamr as a whole now and where it came from. How many people are in the team? For how long have you been working on Streamr?
We are around 35 people with one office in Zug, Switzerland and another one in Helsinki. But there are team members all over the globe, we’ve people in the US, Spain, the UK, Germany, Poland, Australia and Singapore. I joined Streamr back in 2017 during the ICO craze (but not for that reason!)
And did you raise funds so far? If so, how did you handle them? Are you planning to do any future raises?
We did an ICO back in Sept/Oct 2017 in which we raised around 30 Millions CHF. The funds give us enough runway for around five/six years to finalize our roadmap. We’ve also simultaneously opened up a sister company consultancy business, TX which helps enterprise clients implementing the Streamr stack. We've got no more plans to raise more!
What is the token use case? How did you make sure it captures the value of the ecosystem you're building
The token is used for payments on the Marketplace (such as for Data Union products for example) also for the broker nodes in the Network. ( we haven't talked much about the P2P network but it's our project's secret sauce).
The broker nodes will be paid in DATAcoin for providing bandwidth. We are currently working together with Blockscience on our tokeneconomics. We’ve just started the second phase in their consultancy process and will be soon able to share more on the Streamr Network’s tokeneconoimcs.
But if you want to summate the Network in a sentence or two - imagine the Bittorrent network being run by nodes who get paid to do so. Except that instead of passing around static files, it's realtime data streams.
That of course means it's really well suited for the IoT economy.
Well, let's continue with questions from Twitter and this one comes at the perfect time. Can Streamr Network be used to transfer data from IOT devices? Is the network bandwidth sufficient? How is it possible to monetize the received data from a huge number of IOT devices? From u/ EgorCypto
Yes, IoT devices are a perfect use case for the Network. When it comes to the network’s bandwidth and speed - the Streamr team just recently did extensive research to find out how well the network scales.
The result was that it is on par with centralized solutions. We ran experiments with network sizes between 32 to 2048 nodes and in the largest network of 2048 nodes, 99% of deliveries happened within 362 ms globally.
To put these results in context, PubNub, a centralized message brokering service, promises to deliver messages within 250 ms — and that’s a centralized service! So we're super happy with those results.
Here's a link to the paper:
https://medium.com/streamrblog/streamr-network-performance-and-scalability-whitepaper-adb461edd002
While we're on the technical side, second question from Twitter: Can you be sure that valuable data is safe and not shared with service providers? Are you using any encryption methods? From u/ CryptoMatvey
Yes, the messages in the Network are encrypted. Currently all nodes are still run by the Streamr team. This will change in the Brubeck release - our last milestone on the roadmap - when end-to-end encryption is added. This release adds end-to-end encryption and automatic key exchange mechanisms, ensuring that node operators can not access any confidential data.
If BTW - you want to get very technical the encryption algorithms we are using are: AES (AES-256-CTR) for encryption of data payloads, RSA (PKCS #1) for securely exchanging the AES keys and ECDSA (secp256k1) for data signing (same as Bitcoin and Ethereum).
Last question from Twitter, less technical now :) In their AMA ad, they say that Streamr has three unions, Swash, Tracey and MyDiem. Why does Tracey help fisherfolk in the Philippines monetize their catch data? Do they only work with this country or do they plan to expand? From u/ alej_pacedo
So yes, Tracey is one of the first Data Unions on top of the Streamr stack. Currently we are working together with the WWF-Philippines and the UnionBank of the Philippines on doing a first pilot with local fishing communities in the Philippines.
WWF is interested in the catch data to protect wildlife and make sure that no overfishing happens. And at the same time the fisherfolk are incentivized to record their catch data by being able to access micro loans from banks, which in turn helps them make their business more profitable.
So far, we have lots of interest from other places in South East Asia which would like to use Tracey, too. In fact TX have already had explicit interest in building out the use cases in other countries and not just for sea-food tracking, but also for many other agricultural products.
(I think they had a call this week about a use case involving cows 😂)
I recall late last year, that the Streamr Data Union framework was launched into private beta, now public beta was recently released. What are the differences? Any added new features? By u/ Idee02
The main difference will be that the DU 2.0 release will be more reliable and also more transparent since the sidechain we are using for micropayments is also now based on blockchain consensus (PoA).
Are there plans in the pipeline for Streamr to focus on the consumer-facing products themselves or will the emphasis be on the further development of the underlying engine?by u/ Andromedamin
We're all about what's under the hood. We want third party devs to take on the challenge of building the consumer facing apps. We know it would be foolish to try and do it all!
As a project how do you consider the progress of the project to fully developed (in % of progress plz) by u/ Hash2T
We're about 60% through I reckon!
What tools does Streamr offer developers so that they can create their own DApps and monetize data?What is Streamr Architecture? How do the Ethereum blockchain and the Streamr network and Streamr Core applications interact? By u/ CryptoDurden
We'll be releasing the Data UNion framework in a few weeks from now and I think DApp builders will be impressed with what they find.
We all know that Blockchain has many disadvantages as well,
So why did Streamr choose blockchain as a combination for its technology?
What's your plan to merge Blockchain with your technologies to make it safer and more convenient for your users? By u/ noonecanstopme
So we're not a blockchain ourselves - that's important to note. The P2P network only uses BC tech for the payments. Why on earth for example would you want to store every single piece of info on a blockchain. You should only store what you want to store. And that should probably happen off chain.
So we think we got the mix right there.
What were the requirements needed for node setup ? by u/ John097
Good q - we're still working on that but those specs will be out in the next release.
How does the STREAMR team ensure good data is entered into the blockchain by participants? By u/ kartika84
Another great Q there! From the product buying end, this will be done by reputation. But ensuring the quality of the data as it passes through the network - if that is what you also mean - is all about getting the architecture right. In a decentralised network, that's not easy as data points in streams have to arrive in the right order. It's one of the biggest challenges but we think we're solving it in a really decentralised way.
What are the requirements for integrating applications with Data Union? What role does the DATA token play in this case? By u/ JP_Morgan_Chase
There are no specific requirements as such, just that your application needs to generate some kind of real-time data. Data Union members and administrators are both paid in DATA by data buyers coming from the Streamr marketplace.
Regarding security and legality, how does STREAMR guarantee that the data uploaded by a given user belongs to him and he can monetize and capitalize on it? By u/ kherrera22
So that's a sort of million dollar question for anyone involved in a digital industry. Within our system there are ways of ensuring that but in the end the negotiation of data licensing will still, in many ways be done human to human and via legal licenses rather than smart contracts. at least when it comes to sizeable data products. There are more answers to this but it's a long one!
Okay thank you all for all of those!
The AMA took place in the GAINS Telegram group 10/09/20. Answers by Shiv Malik.
submitted by thamilton5 to streamr [link] [comments]

ElectrumX vs EPS and security of the server

Hi team, can you please explain two things:
1) I have read there is also an implementation called ElectrumX. But I only found a guide how to installl EPS by Chris Belcher on github. Can you linke me to a nice guide for ElectrumX? What are the main differences between ElectrumX vs EPS ?
2) What security related tradeoffs are you making when using EPS for your full node? EPS is a server, so are not servers able to "spy" on server activities and see everything in an un-encrypted manner?
3) When you run your own EPS how can you make sure that it works also in years? From what is its functionality depending? Is it some sort decentralized mash network? I like to know which scenario must take place in order for the EPS server infrastructure not to work anymore? the Worst case
Thanks all
submitted by celentano1234 to Electrum [link] [comments]

08-21 10:34 - 'I finally finished an altcoin with working mining, wallet etc. However when I tried to run a main node on AWS I could not for the life of me get my local client to connect.' (self.Bitcoin) by /u/AthP2121 removed from /r/Bitcoin within 67-77min

'''
I opened all ports on AWS, put in the DNS and public address under “addnode” yet still nothing. Is this a common issue with AWS?
(Please let me know if you want any more details on the project)
'''
I finally finished an altcoin with working mining, wallet etc. However when I tried to run a main node on AWS I could not for the life of me get my local client to connect.
Go1dfish undelete link
unreddit undelete link
Author: AthP2121
submitted by removalbot to removalbot [link] [comments]

RiB Newsletter #14 – Are We Smart (Contract) Yet?

We’re seeing a bunch of interesting Rust blockchain and crypto projects, so this month the “Interesting Things” section is loaded up with news, papers, and project links.
This month, Elrond, appeared on our radar with the launch of their mainnet. Although not written in Rust, it runs Rust smart contracts on its Arwen WASM VM, which itself is based on the Rust Wasmer VM. Along with NEAR, Nervos, and Enigma (and probably others), this continues an encouraging trend of blockchains enabling smart contracts in Rust. See the “Interesting Things” section for examples of Elrond’s Rust contracts.
Rust continues to be popular for research into zero-knowledge proofs, with Microsoft releasing Spartan, a zk-SNARK system without trusted setup.
In RiB news, we published a late one-year anniversary blog post. It has some reflection on the changes to, and growth of, RiB over the last year.
The Awesome Blockchain Rust project, which is maintained by Sun under the rust-in-blockchain GitHub org, has received a stream of updates recently, and is now published as the Awesome-RiB page on rustinblockchain.org.
It’s a pretty good resource for finding blockchain-related Rust projects, with links to many of the more prominent and mature projects noted in the RiB newsletter. It could use more eyes on it though.

Project Spotlight

Each month we like to shine a light on a notable Rust blockchain project. This month that project is…
ethers.rs
ethers.rs is an Ethereum & Celo library and wallet implementation, implemented as a port of the ethers.js library to Rust.
Ethereum client programming is usually done in JavaScript with either web3.js or ethers.js, with ethers.js being the newer of the two. These clients communicate to an Ethereum node, typically via JSON-RPC (or, when in the browser, via an “injected” client provider that follows EIP-1193, like MetaMask).
ethers.rs then provides a strongly-typed alternative for writing software that interacts with the Ethereum network.
As of now it is only suited for non-browser use cases, but if you prefer hacking in Rust to JavaScript, as some of us surely do, it is worth looking into for your next Ethereum project.
The author of ethers.rs, Georgios Konstantopoulos, accepts donations to sponsor their work.
Note that there is also a Rust alternative to web3.js, rust-web3.

Interesting Things

News

Blog Posts

Papers

Projects

Podcasts and Videos


Read more: https://rustinblockchain.org/newsletters/2020-08-05-are-we-smart-contract-yet/
submitted by Aimeedeer to rust [link] [comments]

Air-gapped z-addresses; Is ZecWallet an option?

I've heard of Zcash for a while, but it wasn't until recently that I tried my hand playing around with the daemon and wallets.
Obviously, there's no point in using ZEC if you're only using t-addresses, but my desire for a z-address capable wallet certainly narrows the choice of wallets available to me.
Running a full node is no problem for me; I'd like to take advantage of a GUI if possible though. For that reason, I am drawn to ZecWallet's full node version. But I'm still uneasy when it comes to key security. (Can anyone share their experience with the ZecWallet paper wallet generator?)
From what I gather there is no wallet with HD support for t-addresses, is that right? Not much of a concern for me because I am interested in the shielded pool. I just figured t-addresses would support Electrum-style seeds but apparently not?
Sapling addresses seem to be exactly what I want; in particular I am drawn to their reusability and ability to export the view key. I was hoping this would ease the process of securing and backing up my private keys.
Here's my key handling protocol I use for Monero:
1) Generate the wallet on an air-gapped machine
2) This gives you a mnemonic seed. I write that down and keep it as an analog backup. By using a passphrase in conjunction with the seed, I can effectively encrypt this paper wallet easily.
3) Export the private view key and address to an online machine and make a watch-only wallet. This lets my watching wallet see incoming transactions
4) When outputs are received, I have to export the list of outputs to the air-gapped machine. The air-gapped machine uses this data to make signed key images.
5) I export the key images back to the watching wallet. At this point, the watching wallet can see outgoing transactions.
6) Now I can create unsigned transactions with the watching wallet, sign them in the air-gapped machine, and transmit them via the watching wallet using my full node.
The major benefit of using Monero in this way is that I only have to make a human-readable backup of my wallet once and I'm set for life.
Obviously, Zcash is going to be a little bit different. Since the core client doesn't give us mnemonic seed phrases, that complicates backup a little bit. What's the best way to back up ZEC? If I keep an up-to-date backup of my wallet.dat is that all I need? Is there an option in the wallet to encrypt this backup as well, or do I need to accomplish that externally with the likes of Veracrypt? I must admit the idea of unencrypted wallet data being written to my disk makes me uneasy.
I see that there is an option in zcash-cli to import/export the view key of Sapling addresses. However, I can't see the option to do so in ZecWallet, and when I do so manually via the CLI nothing seems to be reflected in ZecWallet.
Is ZecWallet by its very nature an obligatory hot wallet, or am I missing some functionality in the wallet?
My end goal is to run a ZEC full node on Qubes and hold my coins in z-addresses. Qubes allows me to make virtually air-gapped VMs to greatly simplify key management.
So for example when I use Bitcoin, I have a networked VM that runs a Bitcoind + Electrum Personal Server + Electrum Wallet stack, where I import my master public key. When I need to sign a transaction, I spin up a networkless VM equipped with Electrum and my private keys. Qrexec let's me easily ferry unsigned/signed transactions back and forth between the two VMs. Overall this provides a decent UX with above-average security and privacy. I'd like to port this general setup to Zcash.
To do so, I need a GUI wallet that supports both z-addresses and public/private key splitting. Does such a tool exist? (Can Electrum Personal Server be ported to ZEC?) If not, how can I streamline this process with the CLI?
I'm more familiar with Monero than Bitcoin, so the Zcash/Bitcoin CLIs are still a little foreign to me, though I am not "afraid" of CLI wallets in general. My Cryptonote muscle memory makes me prone to annoying little syntax errors I'd much rather do without.
My plan is to buy ZEC from Coinbase Pro, withdraw to a t-address, and then sweep my coins to a z-address. I want to monitor the balance of both t-addresses and z-addresses (and later send transactions) without ever exposing my private keys to the Internet.
However, it seems like the Zcash CLI is my only viable option for z-address watching wallets. Should I just play around on testnet until I get more familiar, or is there a GUI wallet solution out there that fits my needs? Does anyone have a cheat sheet for doing this via the CLI that could help me along the learning curve?
TL;DR New to Zcash, need advice as it relates to wallet backup, watching wallets, and z-addresses.
Assistance is much appreciated!
Edit: I don't suppose there's a way to use a Trezor Model T with a full node and or z-addresses?
submitted by spirtdica to zec [link] [comments]

Epic Cash AMA Recap with CryptoDiffer Community

CryptoDiffer team Hello, everyone! We are glad to meet here: Max Freeman (@maxfreeman4), Project Lead at Epic Cash Yoga Dude (@Yogadude), PR&Marketing at Epic Cash Xenolink (@Xenolink), Advisor at Epic Cash
Max Freeman Project Lead at Epic Cash Thanks Max, we are excited to be here!
Yoga Dude PR&Marketing at Epic Cash Hello Everyone! Thank you for having us here!
Xenolink Advisor at Epic Cash Thank you to the CryptoDiffer team and CryptoDiffer community for hosting us!
CryptoDiffer team Let`s start from the first introduction question: Q1: Can you introduce yourself to the community? What is your background and how did you join Epic Cash?
Yoga Dude PR&Marketing at Epic Cash
Hello! My background is Marketing and Business Development, I’ve been in crypto since 2011 started with Bitcoin, then Monero in 2014, Ethereum in 2015 and at some point Doge for fun and profit. I joined Epic Cash team in September 2019 handling PR and Marketing.
I saw in Epic Cash what was missing in my previous cryptos — things that were missing in Bitcoin and Monero especially.
Xenolink Advisor at Epic Cash
Hello Cryptodiffer Community, I am not an original co-founder nor am I a developer for the Epic Cash project. I am however a community member that is involved in helping scale this project to higher levels. One of the many beauties of Epic Cash is that every single member in the community has the opportunity to be part of EPIC’s team, it can be from development all the way to content producing. Epic Cash is a community driven project. The true Core Team of Epic Cash is our community. I believe a community that is the Core Team is truly powerful. EPIC Cash has one of the freshest and strongest communities I have seen in quite a while. Which is one of the reasons why I became involved in this project. Epic displayed some of the most self community produced content I have seen in a project. I’m actually a doctor of medicine but in terms of my experience in crypto, I have been involved in the industry since 2012 beginning with mining Litecoin. Since then I have been doing deep dive analysis on different projects, investing, and building a network in crypto that I will utilize to help connect and scale Epic in every way I can. To give some credit to those people in my network that have been a part of helping give Epic exposure, I would like to give a special thanks to u/Tetsugan and u/Saurabhblr. Tetsugan has been doing a lot of work for the Japanese community to penetrate the Japanese market, and Japan has already developed a growing interest in Epic. Daku Sarabh the owner and creator of Crypto Daku Robinhooders, I would like to thank him and his community for giving us one of our first large AMA’s, which he has supported our project early and given us a free AMA. Many more to thank but can’t be disclosed. Also thank you to all the Epic Community leaders, developers, and Content producers!
Max Freeman Project Lead at Epic Cash
I’m Max Freeman, which stands for “Maximum Freedom for Mankind”. I started working on the ideas that would become Epic in 2018. I fell in love with Bitcoin in 2017 but realized that it needs privacy at the base layer, fungibility, better scalability in order to go to the next level.
CryptoDiffer team
Really interesting backgrounds I must admit, pleasure to see the team that clearly has one vision of the project by being completely decentralized:)
Q2: Can you briefly describe what is Epic Cash in 3–5 sentences? What technology stands behind Epic Cash and why it’s better than the existing one?
Max Freeman Project Lead at Epic Cash
I’d like to highlight the differences between Epic and the two highest-valued privacy coin projects, Monero and Zcash. XMR has always-on privacy like Epic does, but at a cost: Its blockchain is over 20x more data intensive than Epic, which limits its possibilities for scalability. Epic’s blockchain is small and light enough to run a full node on cell phones, something that is in our product road map. ZEC by comparison can’t run on low end devices because of its zero knowledge based approach, and only 1% of transactions are fully private. Epic is simply newer, more advanced technology than prior networks thanks to Mimblewimble
We will also add more algorithms to widen the range of hardware that can participate in mining. For example, cell phones and tablets based around ARM chips. Millions of people can mine Epic that can’t mine Bitcoin, and that will help grow the network rapidly.
There are some great short videos on our YouTube channel https://www.youtube.com/channel/UCQBFfksJlM97rgrplLRwNUg/videos
that explain why we believe we have created something truly special here.
Our core architecture derives from Grin, so we are fortunate to benefit on an ongoing basis from their considerable development efforts. We are focused on making our currency truly usable and widely available, beyond a store of value and becoming a true medium of exchange.
Yoga Dude PR&Marketing at Epic Cash
Well we all have our views, but in a nutshell, we offer things that were missing in the previous cryptos. We have sound fiscal emission schedule matching Bitcoin, but we are vastly more private and faster. Our blockchain is lighter than Bitcoin or Monero and our tech is more scalable. Also, we are unique in that we are mineable with CPUs and GPUs as well as ASICs, giving the broadest population the ability to mine Epic Cash. Plus, you can’t forget FUNGIBILITY 🙂 we are big on that — since you can’t have true privacy without fungibility.
Also, please understand, we have HUGE respect to all the cryptos that came before us, we learned a lot from them, and thanks to their mistakes we evolved.
Xenolink Advisor at Epic Cash
To add on, what also makes Epic Cash unique is the ability to decentralize the mining using a tri-algo model of Random X (CPU), Progpow (GPU), and Cuckoo (ASIC) for an ability to do hybrid mining. I believe this is an issue we can see today in Bitcoin having centralized mining and the average user has a costly barrier of entry.
To follow up on this one in my opinion one of the things we adopted that we have seen success for , in example Bitcoin and Monero, is a strong community driven coin. I believe having a community driven coin will provide a more organic atmosphere especially when starting with No ICO, or Premine with a fair distribution model for everyone.
CryptoDiffer team
Q3: What are the major milestones Epic Cash has achieved so far? Maybe you can share with us some exciting plans for future weeks/months?
Yoga Dude PR&Marketing at Epic Cash
Since we went live in September of 2019, we attracted a very large community of users, miners, investors and contributors from across the world. Epic Cash is a very international project with white papers translated into over 30 languages. We are very much a community driven project; this is very evident from our content and the amount of translations in our white papers and in our social media content.
We are constantly working on improving our usability, security and privacy, as well as getting our message and philosophy out into the world to achieve mass adoption. We have a lot of exciting plans for our project, the plan is to make Epic Cash into something that is More than Money.
You can tell I am the Marketing guy since my message is less about the actual tech and more about the usability and use cases for Epic Cash, I think our Team and Community have a great mix of technical, practical, social and fiscal experiences. Since we opened our YouTube channels content for community submissions, we have seen our content translated into Spanish, French, German, Polish, Chinese, Japanese, Arabic, Russian, and other languages
Max Freeman Project Lead at Epic Cash
Our future development roadmap will be published soon and includes 4 tracks:
Usability
Mining
Core Protocol
Ecosystem Development
Core Protocol
Epic Server 2.9.0 — this release improves the difficulty adjustment and is aimed at making block emission closer to the target 60 seconds, particularly reducing the incidence of extremely short and long blocks — Status: In Development (Testing) Anticipated Release: June 2020
Epic Server 3.0.0 — this completes the rebase to Grin 3.0.0 and serves as the prerequisite to some important functional building blocks for the future of the ecosystem. Specifically, sending via Tor (which eliminates the need to open ports), proof of payment (useful for certain dex applications e.g. Bisq), and our native mobile app. Status: In Development (Testing) Anticipated Release: Fall 2020
Non-Interactive Transactions — this will enhance usability by enabling “fire and forget” send-to-address functionality that users are accustomed to from most cryptocurrencies. Status: Drawing Board Anticipated Release: n/a
Scaling Options — when blocks start becoming full, how will we increase capacity? Two obvious options are increasing the block size, as well as a Lightning Network-style Layer 2 structure. Status: Drawing Board Anticipated Release: n/a
Confidential Assets — Similar to Raven, Tari, and Beam, the ability to create independently tradable assets that ride on the Epic Blockchain. Status: Drawing Board Anticipated Release: n/a
Usability
GUI Wallet 2.0 — Restore from seed words and various usability enhancements — Status: Needs Assessment Anticipated Release: Fall 2020
Mobile App — Native mobile experience for iOS and Android. Status: In Development (Testing) Anticipated Release: Winter 2020
Telegram Integration — Anonymous payments over the Telegram network, bot functionality for groups. Status: Drawing Board Anticipated Release: n/a
Mining
RandomX on ARM — Our 4th PoW algorithm, this will enable tablets, cell phones, and low power devices such as Raspberry Pi to participate in mining. Status: Needs Assessment Anticipated Release: n/a
The economics of mining Epic are extremely compelling for countries that have free or extremely cheap electricity, since anyone with an ordinary PC can mine. Individual people around the world can simply run the miner and earn meaningful money (imagine Venezuela for example), something that has not been possible since the very early days of Bitcoin.
Ecosystem Development
Atomic Swaps — Connecting Epic to other blockchains in a trustless way, starting with ETH so that Epic can trade on DeFi infrastructure such as Uniswap, Kyber, etc. Status: Drawing Board Anticipated Release: n/a
Xenolink Advisor at Epic Cash
From the Community aspect, we have been further developing our community international reach. We have been seeing an increase in interest from South America, China, Russia, Japan, Italy, and the Philippines. We are working on targeting more countries. We truly aim to be a decentralized project that is open to everyone worldwide.
CryptoDiffer team
Great, thank you for your answers, we now can move to community questions part!
Cryptodiffer Community
You have 3 mining algorithms, the question is: how do they not compete with each other? Is there any benefit of mining on the GPU and CPU if someone is mining on the ASIC?
Max Freeman Project Lead at Epic Cash
The block selection is deterministic, so that every 100 blocks, 60% are for RandomX (CPU), 38% for ProgPow (GPU), and 2% for Cuckoo (ASIC) — the policy is flexible so that we can have as many algorithms with any percentages we want. The goal is to make the most decentralized and resilient network possible, and with that in mind we are excited to work on enabling tablets and cell phones to mine, since that opens it up to millions of people that otherwise can’t take part.
Cryptodiffer Community
To Run a project smoothly, Funding is very important, From where does the Funding/revenue come from?
Xenolink Advisor at Epic Cash
Yes, early on this was realized and in order to scale a project funds are indeed needed. Epic Cash did not start with any funding and no ICO and was organically genesis mined with no pre-mine. Epic cash is also a nonprofit community driven project similar to Monero. There is no profit-driven entity in the picture. To overcome the revenue issue Epic Cash setup a development fund tax that decreases 1% every year until 2028 when Epic Cash reaches singularity with Bitcoin emissions. Currently it is at 7.77%. This will help support the scaling of the project.
Cryptodiffer Community
Hi! In your experience working also with MONERO can you please clarify which are those identified problems that EPIC CASH aims to develop and resolve? What’s the main advantage that EPIC CASH has over MONERO? Thank you!
Yoga Dude PR&Marketing at Epic Cash
First, I must admit that I am still a huge fan and HODLer of Monero. That said:
✅ our blockchain is MUCH lighter than Monero’s
✅ our transaction processing speed is much faster
✅ our address-less blockchain is more private
✅ Epic Cash can be mined with CPU (RandomX) GPU (ProgPow) and Cuckoo, whereas Monero migrated to RandomX and currently only mineable with CPU
Cryptodiffer Community
  1. the feature ‘Cut Through’ deletes old data, how is it decided which data will be deletes, and what are the consequences of it for the platform and therefore the users?
  2. On your website I see links to download Epic wallet and mining software for Linux,Windows and MacOs, I am a user of android, is there a version for me, or does it have a release date?
Max Freeman Project Lead at Epic Cash
  1. This is one of the most exciting features of Mimblewimble, which is its extraordinary ability to compress blockchain data. In Bitcoin, the entire history of a coin must be replayed every time it is spent, and comprehensive details are permanently stored in the blockchain. Epic discards spent transaction inputs and consolidates outputs, storing neither addresses or amounts, only a tiny kernel to allow sender and receiver to prove their transaction.
  2. The Vitex mobile app is great for today, and we have a native mobile app for iOS and Android in the works as well.
Cryptodiffer Community
$EPIC Have total Supply of 21,000,000 EPIC , is there any burning plan? Or Buyback program to maintain $EPIC price in the future?
Who is Epic Biggest competitors?
And what’s makes epic better than competitors?
Xenolink Advisor at Epic Cash
We respect the older generation coins like Bitcoin. But we have learned that the supply economics of Bitcoin is very sound. Until today we can witness how the Bitcoin is being adopted institutionally and by retail. We match the 21 million BTC supply economics because it is an inelastic fixed model which makes the long-term economics very sound. To have an elastic model of burning tokens or printing tokens will not have a solid economic future. Take for example the USD which is an inflating supply. In terms of competitors we look at everyone in crypto with respect and also learn from everyone. If we had to compare to other Mimblewimble tech coins, Grin is an inelastic forever inflating supply which in the long term is not sound economics. Beam however is an inelastic model but is formed as a corporation. The fair distribution is not there because of the permanent revenue model setup for them. Epic Cash a non-profit development tax fund model for scaling purposes that will disappear by 2028’s singularity.
Cryptodiffer Community
What your plans in place for global expansion, are you focusing on only market at this time? Or focus on building and developing or getting customers and users, or partnerships?
Yoga Dude PR&Marketing at Epic Cash
Since we are a community project, we have many developers, in addition to the core team.
Our plans for Global expansion are simple — we have advocates in different regions addressing their audiences in their native languages. We are growing organically, by explaining our ideology and usability. The idea is to grow beyond needing a fiat bridge for crypto use, but to rather replace fiat with our borderless, private and fungible crypto so people can use it to get goods and services without using banks.
We are not limiting ourselves to one particular demographic — Epic Cash is a valid solution for the gamers, investors, techie and non techie people, and the unbanked.
Cryptodiffer Community
EPIC confidential coin! Did you have any problems with the regulators? And there will be no problems with listing on centralized exchanges?
Xenolink Advisor at Epic Cash
In terms of structure, we are carefully set up to minimize these concerns. Without a company or investors in the picture, and having raised no funds, there is little scope to attack in terms of securities laws. Bitcoin and Ethereum are widely acknowledged as acceptable, and we follow in their well-established footprints in that respect. Centralized exchanges already trade other privacy coins, so we don’t see this as much of an issue either. In general, decentralized p2p exchange options are more interesting than today’s centralized platforms. They are more censorship resistant, secure, and privacy-protecting. As the technology gets better, they should continue to gain market share and that’s why we’re proud to be partnered with Vitex, whose exchange and mobile app work very well.
Cryptodiffer Community
What are the main utility and real-life usage of the #EPIC As an investor, why should we invest in the #EPIC project as a long-term investment?
Max Freeman Project Lead at Epic Cash
Because our blockchain is so light (only 1.16gb currently, and grows very slowly) it is naturally well suited to become a decentralized mobile money standard because people can run a full node on their phone, guaranteeing the security of their funds. Scalability in Bitcoin requires complicated and compromised workarounds such as Lightning Network and light clients, and these problems are solved in Epic.
With our forthcoming Mobile Mining app, hundreds of millions of cell phones and tablets will be able to easily join the network. People can quickly and cheaply send money to one another, fulfilling the long-envisioned promise of P2P electronic cash.
As an investor, it’s important to ask a few key questions. Bitcoin Standard tokenomics of disinflation and a fixed supply are well proven over a decade now. We follow this model exactly, with a permanently synchronized supply from 2028, and 4 emission halvings from now until then, with our first one in about two weeks. Beyond that, we can apply some simple logical tests. What is more valuable, money that can only be used in some cases (censorable Bitcoin based on a lack of fungibility) or money that can be used universally? (fungible Epic based on always-on privacy by default). Epic is also poised to be a more decentralized and therefore resilient network because of wider participation in mining. Epic is designed to be Bitcoin++ Privacy, Fungibility, Scalability
Cryptodiffer Community
Q1. What are advantages for choosing three mining algorithms RandomX+, ProgPow and CuckAToo31+ ?
Q2. Beam and Grin use MimbleWimble protocol, so what are difference for Epic? All of you will be friends for partners or competitors?
Max Freeman Project Lead at Epic Cash
RandomX and ProgPow are designed to use the entirety of a CPU / GPU’s unique processing capabilities in a way that other types of hardware don’t work as well. You can run RandomX on a GPU but it doesn’t work nearly as well as a much cheaper CPU, for example. Cuckoo is a “memory hard” algorithm that widens the range of companies that can produce the hardware.
Grin and Beam are great projects and we’ve learned a lot from them. We inherited our first codebase from Grin’s excellent Rust design, which is a better language for community participation than C++ that Beam currently uses.
Functionally, Mimblewimble is similar across the 3 coins, with standard Confidential Transactions, CoinJoin, Dandelion++, Schnorr Signatures and other advanced features. Grin is primarily ASIC-targeted, Beam is GPU-targeted, and Epic is multi-hardware.
The biggest differences though are in tokenomics and project structure. Grin has permanent inflation of 60 coins per block with no halvings, which means steady erosion of value over time due to new supply pressure. It also lacks a steady funding model, making future development in jeopardy, particularly as the per coin price falls. Beam has a for-profit model with heavy early inflation and a high developer tax. Epic builds on the strengths of these earlier mimblewimble projects and addresses the parts that could be improved.
Cryptodiffer Community Some privacy coin has scalability issues! How Epic cash will solve scalability issues? Why you choose randomX consensus algorithem?
Xenolink Advisor at Epic Cash
Fungibility means that you can’t distinguish one unit of currency from another, in example Gold. Fungibility has recently become a hot issue as people have been noticing Bitcoins being locked up by exchanges which may of had a nefarious history which are called Tainted Coins. In example coins that have been involved in a hack, darknet market transactions, or even processing coin through a mixer. Today we can already see freshly mined Bitcoins being sold at a premium price to avoid the fungibility problem Bitcoin carries today. Bitcoin can be tracked by chainalysis and is not a fungible cryptocurrency. One of the features that Epic has is privacy with added fungibility, because of Mimblewimble technology, Epic has no addresses recorded and therefore nothing can be tracked by chainalysis. Below I provide a link of an example of what the lack of fungibility is resulting in today with Bitcoin. One of the reasons why we chose the Random X algo. is because of the easy barrier of entry and also to further decentralize the mining. Random X algo can be mined on old computers or laptops. We also have 2 other algos Progpow (GPU), and Cuckoo (ASIC) to create a wider decentralization of mining methods for Epic.
Cryptodiffer Community
I’m a newbie in crypto and blockchain so how will Epic Cash team target and educate people who don’t know about blockchain and crypto?
What is the uniqueness of Epic Cash that cannot be found in other project that´s been released so far ?
Yoga Dude Pr&Marketing at Epic Cash
Actually, while we have our white paper translated into over 30 languages, we are more focused on explaining our uses and advantages rather than cold specs. Our tech is solid, but we not get hung up on pure tech talk which most casual users do not need to or care to understand. As long as our fundamentals and tech are secure and user friendly our primary goal is to educate about use cases and market potential.
The uniqueness of Epic Cash is its amalgamation of “whats good” in other cryptos. We use Mimblewimble for privacy and anonymity. Our blockchain is much lighter than our competitors. We are the only Mimblewimble crypto to use a unique cocktail of mining algorithms allowing to be mined by casual miners with gaming rigs and laptops, while remaining friendly to GPU and CPU farmers.
The “uniqueness” is learning from the mistakes of those who came before us, we evolved and learned, which is why our privacy is better, we are faster, we are fungible, we offer diverse mining and so on. We are the best blend — thats powerful and unique
Cryptodiffer Community
Can you share EPIC’s vision for decentralized finance (DEFI)? What features do EPIC have to support DEFI?
Yoga Dude PR&Marketing at Epic Cash
We view Epic as ideally suited to be the decentralized digital reserve asset of the new Private Internet of Money that’s emerging. At a technology level, atomic swaps can be created to build liquidity bridges so that wrapped Epic tokens (like WBTC, WETH) can trade on other networks as ERC20, BEP2, NEP5, VIP180, Algorand and so on. There is more Bitcoin value locked on Ethereum than in Lightning Network, so we will similarly integrate Epic so that it can trade on networks such as Uniswap, Kyber, and so on.
Longer term, if there is market demand for it, thanks to Scriptless Script functionality our blockchain has, we can build “Confidential Assets” (which Raven, Tari, and Beam are all also working on) that enable people to create tokenized assets in a private way.
Cryptodiffer Community
If you could choose one celebrity to promote Epic-cash, who that would be?
Max Freeman Project Lead at Epic Cash
I am a firm believer that the strength of the project lies in allowing community members to become their own celebrities, if their content is good enough the community will propel them to celebrity status. Organic celebrities with small but loyal following are vastly more beneficial than big name professional shills with inflated but non caring audiences.
I remember the early days of Apple when an enthusiastic dude named Guy Kawasaki became Apple Evangelist, he was literally going around stores that sold Apple and visited user groups and Evangelized his belief in Apple. This guy became a Legend and helped Apple become what it is today.
Epic Cash will have its OWN Celebrities
Cryptodiffer Community
How does $EPIC solve scalability of transactions? Current blockchains face issues with scalability a lot, how does $EPIC creates a solution to it?
Xenolink Advisor at Epic Cash
Epic Cash is utilizing Mimblewimble technology. Besides the privacy & fungibility aspect of the tech. There is the scalability features of it. It is implemented into Epic by transaction cut-through. Which means it allows nodes to remove all intermediate transactions, thus significantly reducing the blockchain size without affecting its validation. Mimblewimble also does not use addresses like a BTC address, and amount of transactions are also not recorded. One problem Monero and Bitcoin are facing now is scalability. It is evident today that data is getting more expensive and that will be a problem in the long run for those coins. Epic is 90% lighter and more scalable compared to Monero and Bitcoin.
Cryptodiffer Community
what are the ways that Epic Cash generates profits/revenue to maintain your project and what is its revenue model ? How can it make benefit win-win to both invester and your project ?
Max Freeman Project Lead at Epic Cash
There is a block subsidy of 7.77% that declines 1.11% per year until 0, where it stays after that. As a nonprofit community effort, this extremely modest amount goes much further than in other projects, which often take 20, 30, even 50+ % of the coin supply. We believe that this ongoing funding model best aligns the long term incentives for all participants and balances the compromises between the ends of the centralized/decentralized spectrum of choices that any project must make.
Cryptodiffer Community
Q1 : What are your major goals to archive in the next 3–4 years?
Q2 : What are your plans to expand and gain more adoption?
Yoga Dude Pr&Marketing at Epic Cash
Max already talked about our technical plans and goals in his roadmap. Allow me to talk more about the non technical 😁
We are aiming for broader reach in the non technical more mainstream community — this is a big challenge but we believe it is doable. By offering simpler ways to mine Epic Cash (with smart phones for example), and by doing more education we will achieve the holy grail of crypto — moving past the fiat bridges and getting Epic Cash to be accepted as means of payment for goods and services. We will accomplish this by working with regional advocacy groups, community interaction, off-line promotional activities and diverse social media targeting.
Cryptodiffer Community
It seems to me that EpicCash will have its first Halving, right? Why a halving so soon?
Is a mobile version feasible?
Max Freeman Project Lead at Epic Cash
Our supply emission catches up to that of Bitcoin’s first 19 years after 8 years in Epic, so that requires more frequent halvings. Today’s block emission is 16, next up are 8, 4, 2, and then finally 0.15625. After that, the supply of Epic and that of BTC stay synchronized until maxing out at 21m coins in 2140.
Today we have a mobile wallet through the Vitex app, a native mobile wallet coming, and are working on mobile mining.
Cryptodiffer Community
What markets will you add after that?
Yoga Dude PR&Marketing at Epic Cash
Well, we are aiming to have ALL markets
Epic Cash in its final iteration will be usable by everyone everywhere regardless of their technical expertise. We are not limiting ourselves to the technocrats, one of our main goals is to help the billions of unbanked. We want everyone to be able to mine, buy, and most of all USE Epic Cash — gamers, farmers, soccer moms, students, retirees, everyone really — even bankers (well once we defeat the banking industry)
We will continue building on the multilingual diversity of our global community adding support and advocacy groups in more countries in more languages.
Epic Cash is More than Money and its for Everyone.
Cryptodiffer Community
Almost, all cryptocurrencies are decentralized & no-one knows who owns that cryptocurrencies ! then also, why Privacy is needed? hats the advantages of Private coins?
Max Freeman Project Lead at Epic Cash
With a public transparent blockchain such as Bitcoin, you are permanently posting a detailed history of your money movements open for anyone to see (not just legitimate authorities, either!) — It would be considered crazy to post your credit card or bank statements to Twitter, but that’s what is happening every time you send a transaction that is not private. This excellent video from community contributor Spencer Lambert https://www.youtube.com/watch?v=0blbfmvCq\_4 explains better than I can.
Privacy is not just for criminals, it’s for everyone. Do you want your landlord to increase the rent when he sees that you get a raise? Your insurance company to raise your healthcare costs because they see you buying too much ice cream? If you’re a business, do you want your employees to see how much money their coworkers make? Do you want your competitors to trace your supplier and customer relationships? Of course not. By privacy being default for everyone, cryptocurrency can be used in a much wider range of situations without unacceptable compromises.
Cryptodiffer Community
What are the main utility and real-life usage of the #EPIC As an investor, why should we invest in the #EPIC project as a long-term investment?
Xenolink Advisor at Epic Cash
Epic Cash can be used as a Private and Fungible store of value, medium of exchange, and unit of account. As Epic Cash grows and becomes adopted it can be compared to how Bitcoin and Monero is used and adopted as well. As Epic is adopted by the masses, it can be accepted as a medium of exchange for store owners and as fungible payments without the worry of having money that is tainted. Epic Cash as a store of value may be a good long term aspect of investment to consider. Epic Cash carries an inelastic fixed supply economic model of 21 million coins. There will be 5 halvings which this month of June will be our first halving of epic. From a block reward of 16 Epic reduced to 8. If we look at BTC’s price action and history of their halvings it has been proven and show that there has been an increase in value due to the scarcity and from halvings a reduction of # of BTC’s mined per block. An inelastic supply model like Bitcoin provides proof of the circulating supply compared to the total supply by the history of it’s Price action which is evident in long term charts since the birth of Bitcoin. EPIC Plans to have 5 halvings before the year 2028 to match the emissions of Bitcoin which we call the singularity event. Below is a chart displaying our halvings model approaching singularity. Once bitcoin and cryptocurrency becomes adopted mainstream, the fungibility problem will be more noticed by the general public. Privacy coins and the features of fungibility/scalability will most likely be sought over. Right now a majority of people believe that all cryptocurrency is fungible. However, that is not true. We can already see Chainalysis confirming that they can trace and track and even for other well-known privacy coins today such as Z-Cash.
Cryptodiffer Community
  1. You aim to reach support from a global community, what are your plans to get spanish speakers involved into Epic Cash? And emerging markets like the african
  2. How am I secure I won’t be affected by receiving tainted money?
Max Freeman Project Lead at Epic Cash
Native speakers from our community are working to raise awareness in key markets such as mining in Argentina and Venezuela for Spanish (Roberto Navarro called Epic “the holy grail of cryptocurrency” and Ethiopia and certain North African countries that have the lowest electricity costs in the world. Remittances between USA and Latin American countries are expensive and slow, so Epic is also perfect for people to send money back home as well.
Cryptodiffer Community
Do EPICs in 2020 focus more on research and coding, or on sales and implementation?
Yoga Dude PR&Marketing at Epic Cash
We will definitely continue to work on research and coding, with emphasis on improved accessibility (especially via smartphones) usability, security and privacy.
In terms of financial infrastructure will continuing to add exchanges both KYC and non KYC.
Big part of our plans is in ongoing Marketing and PR outreach. The idea is to make Epic Cash a viral sensation of sorts. If we can get Epic Cash adopters to spread the word and tell their family, coworkers and friends about Epic Cash — there will be no stopping us and to help that happen we have a growing army of content creators, and supporters.
Everyone with skin in the game gets the benefit of advancing the cause.
Folks also, this isn’t an answer to the question but an example of a real-world Epic Cash content —
https://www.youtube.com/watch?v=XtAVEqKGgqY
a challenge from one of our content creators to beat his 21 pull ups and get 100 epics! This has not been claimed yet — people need to step up 🙂 and to help that I will match another 100 Epic Cash to the first person to beat this
Cryptodiffer Community
I was watching some videos explaining how to send and receive transactions in EpicCash, which consists of ports and sending links, my question is why this is so, which, for now, looks complex?
Let’s talk about the economic model, can EpicCash comply with the concept of value reserve?
Max Freeman Project Lead at Epic Cash
In V3, which is coming later this summer, Epic can be sent over Tor, which eliminates this issue of port opening, even though using tools like ngrok.io, it’s not necessarily as painful as directly configuring the router ports. Early Lightning Network had this issue as well and it’s something we have a plan to address via research into non-interactive transactions. “Fire and Forget” payments to an address, as people are used to in Bitcoin, is coming to Epic and we’re excited to develop functionality that other advanced mimblewimble coins don’t yet have. We are committed to constant improvement in usability and utility, to make our money system the ease of use leader.
We are involved in the project (anyone can join the Freeman Family) because we believe that simply by choosing to use a form of money that better aligns with our ideals, that we can make a positive change in the world. Some of my thoughts about how I got involved are here: https://medium.com/epic-cash/the-freeman-family-e3b9c3b3f166
Max Freeman Project Lead at Epic Cash
Huge thanks to our friends Maks and Vladyslav, we welcome everyone to come say hi at one of our friendly communities. It is extremely early in this journey, our market cap is only 0.5m right now, whereas the 3 other mimblewimble coins are at $20m, $30m and $100m respectively. Epic is a historic opportunity to follow in the footsteps of legends such as Bitcoin and Monero, and we hope to become the first Top 5 privacy coin project.
Xenolink Advisor at Epic Cash
Would like to Thank the Cryptodiffer Team and the Cryptodiffer community for hosting us and also engaging with us to learn more about Epic. If anyone else has more questions and wants to know more about EPIC , can find us at our telegram channel at https://t.me/EpicCash .
Yoga Dude Pr&Marketing at Epic Cash
Thank you, CryptoDiffer Team, and this wonderful Community!!!
Cryptodiffer TEAM
Thank you everyone for taking your time and asking great questions
Thank you for your time, it was an insightful session
Spread the love
submitted by EpicCashFrodo to epiccash [link] [comments]

Port forwarding multiple computers on same network

i am trying to set up bitcoin full node and trying to do so with 2 computers on the same network.
adding first computer went fine, but when trying to add same port with different ip adress i get an error message "the rules cannot be created because multiple rules were specified for a single port / protocol combination"

does this mean i cant have 2 computers acessing the same port on the same network ?
im not really good with this stuff but i try to follow the readme on bitcoin.com and want to run 2 computers
submitted by nelfiweezy to Bitcoin [link] [comments]

CCS idea, maybe someone can write it up - Trans-network peer list sharing - a trans network peer finding system

So, back in the day our CCS used to have a pre CCS ideas section, where manic people (like myself) could pitch ideas, and then someone else could come along and write a CCS to bring that manic idea to fruition. It was called the ideas section. I used to fill it with a lot of ideas. Maybe thats why it was shut down.
Regardless. Here we go.

The problem

All peer to peer networks have an incredibly soft and vulnerable centralized underbelly. In order to find peers, your p2p software needs to get a list of peers.
But where does it get this from?
The current model is blatantly centralized, and everyone just sort of shrugs it off and says "well, you gotta do it somehow". You plop some IP addressed into some DNS records, and then rely on the Domain Name System (DNS) (itself a weird kinda decentralized thing). So now you have a centralized entity (the domain name owner) working through a kinda centralized system (domain ownership isn't entirely permissionless). In addition to this system, there are hardcoded seed peers in the code. Again, someone plops those IP addresses there.
Monero does this. Bitcoin does this. I'm sure any of these networks does this in some fashion. Well, all except for Cryptonote (not our protocol, but an old bitcoin fork that had a cool bifurcated state storage system) - i could never find peers on that network. And the client was buggy.
Anyway. Basically, any time the monero software gets ran for the first time, the software has to phone home to get data it needs to get itself connected to the network.
This is bollucks. A piece of software needs to be able to thrive without depending on external data. A decentralized network needs to be .... well. Decentralized.

The Solution

Trans-network peer list sharing and brute force.
The most obvious solution is that the monero software should just start scanning IP space randomly to find a peer. There's a research paper somewhere about this. If I find it later I'll post it. Surprisingly, the numbers aren't that bad. For a network size of bitcoin, for instance, you only need to randomly scan like 5k IP addresses to statistically find an IP address running a bitcoin node. For monero, its much more, because our network is smaller.
So how do we make this possible? Simple - we work together.
Thus, I propose the trans-network peer list sharing protocol. Basically, we create a piece of software (it could be part of monero, it could be its own thing - who cares, make a library, people turn it into whatever) that basically listens on all ports. It could be all, it could be just the ones you care about. The software hears a request for its master peer list. It then sends that master peer list to the requester. The requester then scans through the list, looking for a node it can peer with.
So basically, I would run a Monero node and this TNPLS software. My monero node would do its thing, and maintain its own peer list. Occasionally, the TNPLS software would pull this list into its database.
My TNPLS is listening for peers. Another TNPLS peer (Bob) connects to me (Alice). Bob says "holy shit, another TNPLS peer!" and Alice is also similarly excited. Bob also runs bitcoin, so it has a trusted bitcoin peer list. But Bob has been running TNPLS for a while, so he also has peer lists from the Litecoin network, edonkey, .... whatever other p2p network is out there. So he puts the bitcoin nodes in their and flags them with trust level 0. He also adds in the litecoin network and edonkey nodes, and they have trust level 1 - because it turns out that he got those peer lists from a TNPLS peer (Sharon) that was running litecoin and edonkey nodes. Bob sends this list to Alice. Alice says thanks very much. Here's my monero node list, with a trust level of 0, because I made them. They my babies.
The two maintain the connection.
So now Alice gets these new bitcoin node peers, and litecoin and edonkey, and puts them in its database. Bitcoin gets a level of 1 (1 degree of separation), and litecoin and edonkey get a 2. It keeps updating its monero nodes. And it occasionally updates the level 1 if Bob is still up and running. And maybe there's some time decay. I dunno. Peer sharing is mostly figured out intranetwork, so I think those mechanisms will more or less work here.
And so now Alice is running her TNPLS peer, listening to anyone sending a TNPLS request on any port. A new TNPLS connects (Fred), running Wownero and looking for nodes. Alice doesn't have any wownero peers in its list (who would?), but Alice tells Fred "Hey buddy, lets see if we can find you some nodes". So Fred sends Alice a little packet that says "Hi my name is Fred and I need a wownero node pls, here is my IP address". Alice then sends that little packet to all of its connected TNPLS peers (just Bob i guess), and those peers look through their peer lists for wownero nodes. Turns out, none of them have anything! So, these nodes send this packet out to all of their connected TNPLS peers (sharon and geraldine). Turns out, Geraldine (geraldine? seriously?) has a wownero node! So Geraldine makes a connection with Fred and says "hey I heard you needed some wownero nodes, here's a list".
Any ddos could be countered with PoW requirements in the protocol, using randomx of course.
edit because i switched identities somewhere in there
submitted by gingeropolous to Monero [link] [comments]

How to keep the last privacy in the era of network transparency

How to keep the last privacy in the era of network transparency

https://preview.redd.it/ngfkz37opv051.jpg?width=1600&format=pjpg&auto=webp&s=896d8941338bc88da6d1199c3e2048a9aaa77c0e
Before half of 2020, the word "data breach" appears extremely active. All over the world are plagued by data breaches, but also cause major losses.
In today's Internet era, any behavior you have on the Internet is likely to be recorded, and then through big data summary and statistical analysis, you can basically say: everything you know, the network knows. In a centralized system, the system platform operator can get all your data in the background. Based on the drive of business interests, they will use this data to commercialize applications: sell data and sell services.
The world is interconnected. This is the status quo and an irreversible development trend. In this interconnection, there are no boundaries in the future. In this near future where there are no borders and everything is connected, imagine that your alarm clock, electricity meter, mobile phone, mobile detector, and other things that are needed every day are interconnected, so that others can understand your situation. What a terrible thing it is, like being in a completely privacy-free environment, such a future, you,Suffocation? Should we have privacy?
In December 1948, the United Nations promulgated the Basic Law, "Universal Declaration of Human Rights," Article 12 of which stated personal privacy as follows:
No one's private life, family, residence and correspondence must be arbitrarily interfered, and his honor and reputation must not be attacked. Everyone has the right to legal protection against such interference or attacks.
In the electronic age, privacy is essential for an open society. Privacy is different from secret. Privacy is something that someone does not want to make public. The secret is something he doesn’t want anyone to know. Privacy is a power. It gives someone the right to decide what to disclose and what not to disclose.
In a distributed Internet environment, the privacy of individuals from a macro perspective mainly covers four aspects: node privacy, content privacy, link privacy, and tunnel privacy. Let's expand one by one and look at the specific content of the four dimensions of privacy.
1) Node privacy
Node privacy refers to the fact that in an open distributed environment, both parties interacting with each other do not know each other's sensitive information, such as IP address and MAC address, so as to achieve the purpose of not exposing each other. At the same time, it is impossible for other nodes to perceive the location of the sender and the receiver through network sniffing.
The following uses the Bitcoin network topology as an example to illustrate the importance of node encryption.

https://preview.redd.it/7bwl3mqayv051.png?width=2368&format=png&auto=webp&s=b3cd22723c7b275d97754051b6a4c7d8a91be553
As of now, there are about 8,000 nodes in the entire Bitcoin network. Based on current technology, the cost of building a parallel sniffing network is very low. Some researchers have done statistics. When the topology sniffing network starts, after about 10 blocks height, it can basically infer the connection topology of the entire network. Coupled with the fixed time interval of gossip message propagation, it is basically possible to infer the general distribution position of the construction nodes of a transaction information, thereby destroying the privacy of the nodes.
2) Privacy of communication content
The privacy of communication content means that the communication content is only visible to both parties of the interaction. No one can intercept the data from the network, or without the authorization of both parties, no one can see the plain text of the communication.
3) Link privacy
Link privacy refers to the connection established by both parties in communication, which is encrypted; No one has the ability to use the link for data transmission without the relevant key. As shown in the figure below, a-> b, b-> c, c-> d, etc.

https://preview.redd.it/h0jg5ivfyv051.png?width=1576&format=png&auto=webp&s=32303d8b296da0cee4e6a4b3a744eb984aa70385
4) Tunnel privacy
In some cases, the communication between the nodes will be completed by one or more relay nodes. In this way, based on the transceiver node and the relay node, a communication tunnel is formed; tunnel privacy means that only the sending node has the right to send data from the sending node to the receiving node via the relay node. Under the premise of authorization, there is no way to complete the transmission of data. As shown above, a-> d, e-> h logical communication tunnel.
For the four privacy dimensions mentioned above, there is a general solution that can effectively protect the security. Although the communication efficiency needs to be improved, functionally speaking, it can already take into account the four dimensions. And in the following, for this general security idea, gives the possible dimensions for further optimization.
1. Description of General Encryption Network Solution
Firstly, each relay routing node in the encrypted network creates a routing descriptor, which contains some contact information, mainly IP addresses, ports, public keys, and other broadband capabilities. After the creation is complete, send this information to the directory server of the whole network (usually also become the Bootstrap node). Based on this information, the directory server generates a unique descriptor for the routing node for the entire network, which is stored on the directory server along with the descriptor information. In the following, we will describe in detail how the privacy of the encrypted network is protected from three aspects: networking topology, message structure and link transmission construction.
1) Network topology
As shown in the following figure, in an encrypted network, we recommend that clients, relay agents, relay routing nodes, directory servers, and possibly bridge nodes together form the entire network topology. When the client builds a communication link, the steps are as follows:
A. The client initiates a node request to the directory server;
B. The directory server generally recommends three nodes to the client from the directory table based on the weight selection algorithm. Logically, they are called entrance node, intermediate node and exit node.
C. After receiving effective feedback from the directory server, the client builds a complete tunnel link step by step according to the Response message.

https://preview.redd.it/dhkgbsfnyv051.png?width=2382&format=png&auto=webp&s=150cfecf95246dccca10ea8a75c904938731a2d0
2) Link establishment
Based on the above description, we know that a client can obtain three nodes of a link through the directory server: entrance node, intermediate node, and exit node.
A. The client uses the DH handshake protocol (Diffie-Hellman) to shake hands with the ingress node to generate a shared session key. Based on the shared key, the client sends a CREATE message to the entrance node;
B. After receiving the CREATE message, the entrance node will establish a link with the intermediate node based on the address of the intermediate node in the message and complete the key exchange;
C. Based on the segmented encrypted link and DH handshake protocol completed above, the client completes key negotiation with the intermediate node;
D. Similarly, based on the two-level encrypted tunnel established above, the client sends a CREATE message to the intermediate node to complete the establishment of the encrypted link between the intermediate node and the exit node;
E. Finally, the client completes the key negotiation between the client and the egress node based on the above three-level segmented encrypted link, and then completes the establishment of the entire onion tunnel link;

https://preview.redd.it/e8mmnz3syv051.png?width=911&format=png&auto=webp&s=de5f22a8c8b3a33ab5ff9274e2145a392b83ca04
3) Message structure and transmission
Before the message is sent from the client, it will use the shared key negotiated with the exit node, intermediate node, and entrance node to encrypt from the inside out. The innermost message is encrypted using the shared key of the exit node, then the intermediate node, and finally the entrance node.

https://preview.redd.it/aiondkuuyv051.png?width=865&format=png&auto=webp&s=a1d3a6877fbee82f668ca3fbea477b86b8b14c8a
A. After the above message is sent from the client, the entrance node will judge the validity of the message based on the shared key negotiated and remove the outer encryption, and then send it to the intermediate node;
B. After receiving the message from the entrance node, the intermediate node will judge the validity of this message based on the shared key negotiated with the client and remove the encryption of this layer, and then send it to the exit node;
C. The exit node uses the shared key negotiated with the client and repeats the above steps. Eventually send the client's message to the real destination address.
2. Optimization for the above program
In the description of the above scheme, we can easily see that there are two obvious flaws, that is, the startup node is too centralized, which can easily lead to a single point of failure or suffer from a network hijacking attack. At the same time, because the data exchange is based on link exchange, when the network congestion is severe, it is easy to cause network service delay. Due to the single structure of the message, it also restricts the data in the link transmission process to a certain extent, and optimizes the transmission performance. Below we will give specific optimization ideas based on the above two points:
1) No central server
The aforementioned encrypted network node knows the existence of all relay and entry / exit nodes by connecting to the directory server. The optimized project node will know the existence of other nodes through the local network database (tentatively called NetDB). NetDB learns the existence of more nodes when connecting other nodes through the DHT algorithm. It is a distributed network database. It mainly provides router contact information and target contact information. Each piece of data is signed by the appropriate party and verified by anyone who uses or stores it.
2) Optimization of data exchange mode
The first point to note is that in the optimized encrypted network, there are two different links for the communication link of the two parties, that is, the entrance link and the exit link are different;

https://preview.redd.it/b7es3d2yyv051.png?width=888&format=png&auto=webp&s=28da350509807d6b5a1ba60f5fb348ff5fe41ea6
In the optimized encrypted network, the connection is broken up into data packets by the message mechanism (Message), after being cross-transmitted through different TCP or UDP tunnels, the receiver reassembles into a data stream, that is, the optimized encrypted network is based on Packet switching, packet switching can drive some implicit load balancing and help avoid congestion and service interruption.
Undoubtedly, open data sharing is the source of power for the development of data-related industries, but the existing data storage methods and network protocols have many shortcomings. Solving such problems has become a very important step on the road to the next generation of the Internet world. DSP Labs has always kept thinking and exploring the next generation Internet infrastructure. I believe that in the near future, DSP Labs can bring a new choice to the Internet world.
Find us:
Facebook: https://www.facebook.com/DSP-Labs-110430797334480
Twitter: https://twitter.com/DSP_Labs
Telegram: t.me/DSP_Official
Wechat: DSPLabs
submitted by DSP-Lab to u/DSP-Lab [link] [comments]

Bitcoin, Blockchain and Cryptocurrency News for Today (September 14) How to make a DNS seeder for your altcoin Node.js With Passport Authentication  Full Project - YouTube [Part 1]How to make your own bitcoin litecoin dogecoin mining pool wallet configuration Building an Alexa skill that uses data from an external API - Dabble Lab #76

Bitcoin Core ist ein vollständiger Bitcoin-Client und bildet das Rückgrat des Netzwerks. Er bietet ein hohes Maß an Sicherheit, Privatsphäre und Stabilität. Er bietet allerdings weniger Funktionen und benötigt viel Festplatten- und Arbeitsspeicher. Features: Bech32 × Bech32: Bech32 ist ein spezielles Adressformat, dass erst durch SegWit möglich wurde (weitere Informationen finden Sie ... In this tutorial we will build an API using Node.js and Express that retrieves data from Bitcoin Core. This tutorial is meant to be a starter for anyone looking to create Bitcoin applications with… Accordingly, the only thing you need to accept bitcoins is a bitcoin client on the server. It is called bitcoind, it’s just a console version of the client, with all the same familiar functionality. It works through the JSON-RPC protocol, is located under port 8332. All that remains after installation is to set up the client and Node.js. Running a Bitcoin full node comes with certain costs and can expose you to certain risks. This section will explain those costs and risks so you can decide whether you’re able to help the network. Special Cases. Miners, businesses, and privacy-conscious users rely on particular behavior from the full nodes they use, so they will often run their own full nodes and take special safety ... Most Bitcoin nodes also act as a Bitcoin client, which allows transactions to be sent to the network. This means a node acts as your personal interface with the Bitcoin network as a whole. Running a node ensures that your Bitcoin transactions are verified and sent to whoever you’re transacting with. This puts the power to send uncensorable money across the world or across the street in your ...

[index] [26411] [4603] [51539] [15988] [20377] [44676] [8907] [24872] [13445] [10298]

Bitcoin, Blockchain and Cryptocurrency News for Today (September 14)

https://bitcoinexchangeguide.com/bitcoin-blockchain-and-cryptocurrency-news-for-today-september-14th-video CognitiveScale Cortex5 Software to Support Keller ... socks5 proxy bitcoin socks5 proxy bittorrent socks5 proxy brazil socks5 proxy best socks5 proxy blog socks5 proxy checker socks5 proxy chrome socks5 proxy client socks5 proxy chrome extension ... #76: In this tutorial, you'll learn how to create a custom Alexa skill that uses data from an external API. The starting point will be the 'API Starter' skil... Category Science & Technology; Song Now That We Are Free - Orignal 12" Mix; Artist Gladiator; Album Now That We Are Free; Licensed to YouTube by How to pay any invoice with bitcoins and an example of piixpay.com API integration into mobile payment wallet solution.

#