I'd like to run a testnet node for some development work I'm doing (in addition to mainnet node). Can someone tell me how much disk space a full node with and/or without tx index is currently needing? And whether it gets reset periodically or always continues to grow like mainnet? Thank you!
Scalenet and Testnet4 are online and open for business
Over the years, some people have made use of testnet3 to test out scaling performance, and have spammed testnet3 with 32 MB blocks. This has caused testnet3 to get kinda bloated; the blockchain now takes an hour or so to sync, which slows down development. Other people have wanted to do stress testing, but have specifically wanted to avoid inconveniencing other people by spamming testnet3, and have therefore not done so. This slows down development too. To address this issue, I created two new networks: testnet4 and scalenet. Testnet4 is intended to be a low-volume quick-syncing blockchain which is ideal for testing out new transaction formats or applications. It has a 2 MB default block size limit, and comes with aserti3 parameters that make the difficulty recover quickly to CPU-mineable levels. It should remain easy to sync on a low-end VPS or old laptop. Scalenet is intended to be a high-volume blockchain which is ideal for spamming and stress testing software. It comes with a 256 MB initial default block size limit, and uses aserti3 parameters that make it more suitable for accurately simulating mainnet mining difficulty (though it retains the 20-minute difficulty rule). In order to prevent storage costs from getting unreasonable for a testnet, scalenet will be reset every 6-12 months by invalidating the block at height 10,001 and adding a new checkpoint. Scalenet is intended to be feasible to run on a mid-range desktop computer with some strain. Testnet4 and scalenet are now online and essentially complete. The code for both has been merged into BCHN and Electron Cash. Testnet4 has also been successfully synced to by Knuth, Bitcoin Unlimited, and libbitcoincashj. Block explorers for both are available, thanks to Axel Gembe (who runs the code) and sickpig (who wrote the code): http://tbch4.loping.net:3002/ http://sbch.loping.net:3003/ The testnet4 and scalenet MRs were opened on Aug 19th and Aug 27th, and both were merged on Sep 17th. Scalenet reached height 10,000 on October 3rd. Testnet4 and Scalenet support are present in the master branch of BCHN, and will be included in the next release of BCHN. Some other software (e.g. Electron Cash) already has support in their latest release, but most is still pending. See also: https://bitcoincashresearch.org/t/testnet4-and-scalenet/148/7
https://github.com/gridcoin-community/Gridcoin-Research/releases/tag/126.96.36.199 Finally! After over ten months of development and testing, "Fern" has arrived! This is a whopper. 240 pull requests merged. Essentially a complete rewrite that was started with the scraper (the "neural net" rewrite) in "Denise" has now been completed. Practically the ENTIRE Gridcoin specific codebase resting on top of the vanilla Bitcoin/Peercoin/Blackcoin vanilla PoS code has been rewritten. This removes the team requirement at last (see below), although there are many other important improvements besides that. Fern was a monumental undertaking. We had to encode all of the old rules active for the v10 block protocol in new code and ensure that the new code was 100% compatible. This had to be done in such a way as to clear out all of the old spaghetti and ring-fence it with tightly controlled class implementations. We then wrote an entirely new, simplified ruleset for research rewards and reengineered contracts (which includes beacon management, polls, and voting) using properly classed code. The fundamentals of Gridcoin with this release are now on a very sound and maintainable footing, and the developers believe the codebase as updated here will serve as the fundamental basis for Gridcoin's future roadmap. We have been testing this for MONTHS on testnet in various stages. The v10 (legacy) compatibility code has been running on testnet continuously as it was developed to ensure compatibility with existing nodes. During the last few months, we have done two private testnet forks and then the full public testnet testing for v11 code (the new protocol which is what Fern implements). The developers have also been running non-staking "sentinel" nodes on mainnet with this code to verify that the consensus rules are problem-free for the legacy compatibility code on the broader mainnet. We believe this amount of testing is going to result in a smooth rollout. Given the amount of changes in Fern, I am presenting TWO changelogs below. One is high level, which summarizes the most significant changes in the protocol. The second changelog is the detailed one in the usual format, and gives you an inkling of the size of this release.
Note that the protocol changes will not become active until we cross the hard-fork transition height to v11, which has been set at 2053000. Given current average block spacing, this should happen around October 4, about one month from now. Note that to get all of the beacons in the network on the new protocol, we are requiring ALL beacons to be validated. A two week (14 day) grace period is provided by the code, starting at the time of the transition height, for people currently holding a beacon to validate the beacon and prevent it from expiring. That means that EVERY CRUNCHER must advertise and validate their beacon AFTER the v11 transition (around Oct 4th) and BEFORE October 18th (or more precisely, 14 days from the actual date of the v11 transition). If you do not advertise and validate your beacon by this time, your beacon will expire and you will stop earning research rewards until you advertise and validate a new beacon. This process has been made much easier by a brand new beacon "wizard" that helps manage beacon advertisements and renewals. Once a beacon has been validated and is a v11 protocol beacon, the normal 180 day expiration rules apply. Note, however, that the 180 day expiration on research rewards has been removed with the Fern update. This means that while your beacon might expire after 180 days, your earned research rewards will be retained and can be claimed by advertising a beacon with the same CPID and going through the validation process again. In other words, you do not lose any earned research rewards if you do not stake a block within 180 days and keep your beacon up-to-date. The transition height is also when the team requirement will be relaxed for the network.
Besides the beacon wizard, there are a number of improvements to the GUI, including new UI transaction types (and icons) for staking the superblock, sidestake sends, beacon advertisement, voting, poll creation, and transactions with a message. The main screen has been revamped with a better summary section, and better status icons. Several changes under the hood have improved GUI performance. And finally, the diagnostics have been revamped.
The wallet sync speed has been DRASTICALLY improved. A decent machine with a good network connection should be able to sync the entire mainnet blockchain in less than 4 hours. A fast machine with a really fast network connection and a good SSD can do it in about 2.5 hours. One of our goals was to reduce or eliminate the reliance on snapshots for mainnet, and I think we have accomplished that goal with the new sync speed. We have also streamlined the in-memory structures for the blockchain which shaves some memory use. There are so many goodies here it is hard to summarize them all. I would like to thank all of the contributors to this release, but especially thank @cyrossignol, whose incredible contributions formed the backbone of this release. I would also like to pay special thanks to @barton2526, @caraka, and @Quezacoatl1, who tirelessly helped during the testing and polishing phase on testnet with testing and repeated builds for all architectures. The developers are proud to present this release to the community and we believe this represents the starting point for a true renaissance for Gridcoin!
Most significantly, nodes calculate research rewards directly from the magnitudes in EACH superblock between stakes instead of using a two- or three- point average based on a CPID's current magnitude and the magnitude for the CPID when it last staked. For those long-timers in the community, this has been referred to as "Superblock Windows," and was first done in proof-of-concept form by @denravonska.
Network magnitude unit pinned to a static value of 0.25
Max research reward allowed per block raised to 16384 GRC (from 12750 GRC)
New CPIDs begin accruing research rewards from the first superblock that contains the CPID instead of from the time of the beacon advertisement
500 GRC research reward limit for a CPID's first stake
6-month expiration for unclaimed rewards
10-block spacing requirement between research reward claims
Rolling 5-day payment-per-day limit
Legacy tolerances for floating-point error and time drift
The need to include a valid copy of a CPID's magnitude in a claim
10-block emission adjustment interval for the magnitude unit
One-time beacon activation requires that participants temporarily change their usernames to a verification code at one whitelisted BOINC project
Verification codes of pending beacons expire after 3 days
Self-service beacon removal
Burn fee for beacon advertisement increased from 0.00001 GRC to 0.5 GRC
Rain addresses derived from beacon keys instead of a default wallet address
Beacon expiration determined as of the current block instead of the previous block
The ability for developers to remove beacons
The ability to sign research reward claims with non-current but unexpired beacons
As a reminder:
Beacons expire after 6 months pass (180 days)
Beacons can be renewed after 5 months pass (150 days)
Renewed beacons must be signed with the same key as the original beacon
Magnitudes less than 1 include two fractional places
Magnitudes greater than or equal to 1 but less than 10 include one fractional place
A valid superblock must match a scraper convergence
Superblock popularity election mechanics
Yes/no/abstain and single-choice response types (no user-facing support yet)
To create a poll, a maximum of 250 UTXOs for a single address must add up to 100000 GRC. These are selected from the largest downwards.
Burn fee for creating polls scaled by the number of UTXOs claimed
50 GRC for a poll contract
0.001 GRC per claimed UTXO
Burn fee for casting votes scaled by the number of UTXOs claimed
0.01 GRC for a vote contract
0.01 GRC to claim magnitude
0.01 GRC per claimed address
0.001 GRC per claimed UTXO
Maximum length of a poll title: 80 characters
Maximum length of a poll question: 100 characters
Maximum length of a poll discussion website URL: 100 characters
Maximum number of poll choices: 20
Maximum length of a poll choice label: 100 characters
Magnitude, CPID count, and participant count poll weight types
The ability for developers to remove polls and votes
[188.8.131.52] 2020-09-03, mandatory, "Fern"
Backport newer uint256 types from Bitcoin #1570 (@cyrossignol)
Implement project level rain for rainbymagnitude #1580 (@jamescowens)
Upgrade utilities (Update checker and snapshot downloadeapplication) #1576 (@iFoggz)
Provide fees collected in the block by the miner #1601 (@iFoggz)
Add support for generating legacy superblocks from scraper stats #1603 (@cyrossignol)
Port of the Bitcoin Logger to Gridcoin #1600 (@jamescowens)
Implement zapwallettxes #1605 (@jamescowens)
Implements a global event filter to suppress help question mark #1609 (@jamescowens)
Add next target difficulty to RPC output #1615 (@cyrossignol)
Add caching for block hashes to CBlock #1624 (@cyrossignol)
Make toolbars and tray icon red for testnet #1637 (@jamescowens)
Add an rpc call convergencereport #1643 (@jamescowens)
Implement newline filter on config file read in #1645 (@jamescowens)
Implement beacon status icon/button #1646 (@jamescowens)
Add gridcointestnet.png #1649 (@caraka)
Add precision to support magnitudes less than 1 #1651 (@cyrossignol)
Replace research accrual calculations with superblock snapshots #1657 (@cyrossignol)
Publish example gridcoinresearch.conf as a md document to the doc directory #1662 (@jamescowens)
Add options checkbox to disable transaction notifications #1666 (@jamescowens)
Add support for self-service beacon deletion #1695 (@cyrossignol)
Add support for type-specific contract fee amounts #1698 (@cyrossignol)
Add verifiedbeaconreport and pendingbeaconreport #1696 (@jamescowens)
Add preliminary testing option for block v11 height on testnet #1706 (@cyrossignol)
Add verified beacons manifest part to superblock validator #1711 (@cyrossignol)
Implement beacon, vote, and superblock display categories/icons in UI transaction model #1717 (@jamescowens)
Scaling Reddit Community Points with Arbitrum Rollup: a piece of cake
https://preview.redd.it/b80c05tnb9e51.jpg?width=2550&format=pjpg&auto=webp&s=850282c1a3962466ed44f73886dae1c8872d0f31 Submitted for consideration toThe Great Reddit Scaling Bake-Off Baked by the pastry chefs atOffchain Labs Please send questions or comments to [[email protected] ](mailto:[email protected]) 1. Overview We're excited to submit Arbitrum Rollup for consideration to The Great Reddit Scaling Bake-Off. Arbitrum Rollup is the only Ethereum scaling solution that supports arbitrary smart contracts without compromising on Ethereum's security or adding points of centralization. For Reddit, this means that Arbitrum can not only scale the minting and transfer of Community Points, but it can foster a creative ecosystem built around Reddit Community Points enabling points to be used in a wide variety of third party applications. That's right -- you can have your cake and eat it too! Arbitrum Rollup isn't just Ethereum-style. Its Layer 2 transactions are byte-for-byte identical to Ethereum, which means Ethereum users can continue to use their existing addresses and wallets, and Ethereum developers can continue to use their favorite toolchains and development environments out-of-the-box with Arbitrum. Coupling Arbitrum’s tooling-compatibility with its trustless asset interoperability, Reddit not only can scale but can onboard the entire Ethereum community at no cost by giving them the same experience they already know and love (well, certainly know). To benchmark how Arbitrum can scale Reddit Community Points, we launched the Reddit contracts on an Arbitrum Rollup chain. Since Arbitrum provides full Solidity support, we didn't have to rewrite the Reddit contracts or try to mimic their functionality using an unfamiliar paradigm. Nope, none of that. We launched the Reddit contracts unmodified on Arbitrum Rollup complete with support for minting and distributing points. Like every Arbitrum Rollup chain, the chain included a bridge interface in which users can transfer Community Points or any other asset between the L1 and L2 chains. Arbitrum Rollup chains also support dynamic contract loading, which would allow third-party developers to launch custom ecosystem apps that integrate with Community Points on the very same chain that runs the Reddit contracts. 1.1 Why Ethereum Perhaps the most exciting benefit of distributing Community Points using a blockchain is the ability to seamlessly port points to other applications and use them in a wide variety of contexts. Applications may include simple transfers such as a restaurant that allows Redditors to spend points on drinks. Or it may include complex smart contracts -- such as placing Community Points as a wager for a multiparty game or as collateral in a financial contract. The common denominator between all of the fun uses of Reddit points is that it needs a thriving ecosystem of both users and developers, and the Ethereum blockchain is perhaps the only smart contract platform with significant adoption today. While many Layer 1 blockchains boast lower cost or higher throughput than the Ethereum blockchain, more often than not, these attributes mask the reality of little usage, weaker security, or both. Perhaps another platform with significant usage will rise in the future. But today, Ethereum captures the mindshare of the blockchain community, and for Community Points to provide the most utility, the Ethereum blockchain is the natural choice. 1.2 Why Arbitrum While Ethereum's ecosystem is unmatched, the reality is that fees are high and capacity is too low to support the scale of Reddit Community Points. Enter Arbitrum. Arbitrum Rollup provides all of the ecosystem benefits of Ethereum, but with orders of magnitude more capacity and at a fraction of the cost of native Ethereum smart contracts. And most of all, we don't change the experience from users. They continue to use the same wallets, addresses, languages, and tools. Arbitrum Rollup is not the only solution that can scale payments, but it is the only developed solution that can scale both payments and arbitrary smart contracts trustlessly, which means that third party users can build highly scalable add-on apps that can be used without withdrawing money from the Rollup chain. If you believe that Reddit users will want to use their Community Points in smart contracts--and we believe they will--then it makes the most sense to choose a single scaling solution that can support the entire ecosystem, eliminating friction for users. We view being able to run smart contracts in the same scaling solution as fundamentally critical since if there's significant demand in running smart contracts from Reddit's ecosystem, this would be a load on Ethereum and would itself require a scaling solution. Moreover, having different scaling solutions for the minting/distribution/spending of points and for third party apps would be burdensome for users as they'd have to constantly shuffle their Points back and forth. 2. Arbitrum at a glance Arbitrum Rollup has a unique value proposition as it offers a combination of features that no other scaling solution achieves. Here we highlight its core attributes. Decentralized. Arbitrum Rollup is as decentralized as Ethereum. Unlike some other Layer 2 scaling projects, Arbitrum Rollup doesn't have any centralized components or centralized operators who can censor users or delay transactions. Even in non-custodial systems, centralized components provide a risk as the operators are generally incentivized to increase their profit by extracting rent from users often in ways that severely degrade user experience. Even if centralized operators are altruistic, centralized components are subject to hacking, coercion, and potential liability. Massive Scaling. Arbitrum achieves order of magnitude scaling over Ethereum's L1 smart contracts. Our software currently supports 453 transactions-per-second for basic transactions (at 1616 Ethereum gas per tx). We have a lot of room left to optimize (e.g. aggregating signatures), and over the next several months capacity will increase significantly. As described in detail below, Arbitrum can easily support and surpass Reddit's anticipated initial load, and its capacity will continue to improve as Reddit's capacity needs grow. Low cost. The cost of running Arbitrum Rollup is quite low compared to L1 Ethereum and other scaling solutions such as those based on zero-knowledge proofs. Layer 2 fees are low, fixed, and predictable and should not be overly burdensome for Reddit to cover. Nobody needs to use special equipment or high-end machines. Arbitrum requires validators, which is a permissionless role that can be run on any reasonable on-line machine. Although anybody can act as a validator, in order to protect against a “tragedy of the commons” and make sure reputable validators are participating, we support a notion of “invited validators” that are compensated for their costs. In general, users pay (low) fees to cover the invited validators’ costs, but we imagine that Reddit may cover this cost for its users. See more on the costs and validator options below. Ethereum Developer Experience. Not only does Arbitrum support EVM smart contracts, but the developer experience is identical to that of L1 Ethereum contracts and fully compatible with Ethereum tooling. Developers can port existing Solidity apps or write new ones using their favorite and familiar toolchains (e.g. Truffle, Buidler). There are no new languages or coding paradigms to learn. Ethereum wallet compatibility. Just as in Ethereum, Arbitrum users need only hold keys, but do not have to store any coin history or additional data to protect or access their funds. Since Arbitrum transactions are semantically identical to Ethereum L1 transactions, existing Ethereum users can use their existing Ethereum keys with their existing wallet software such as Metamask. Token interoperability. Users can easily transfer their ETH, ERC-20 and ERC-721 tokens between Ethereum and the Arbitrum Rollup chain. As we explain in detail below, it is possible to mint tokens in L2 that can subsequently be withdrawn and recognized by the L1 token contract. Fast finality. Transactions complete with the same finality time as Ethereum L1 (and it's possible to get faster finality guarantees by trading away trust assumptions; see the Arbitrum Rollup whitepaper for details). Non-custodial. Arbitrum Rollup is a non-custodial scaling solution, so users control their funds/points and neither Reddit nor anyone else can ever access or revoke points held by users. Censorship Resistant. Since it's completely decentralized, and the Arbitrum protocol guarantees progress trustlessly, Arbitrum Rollup is just as censorship-proof as Ethereum. Block explorer. The Arbitrum Rollup block explorer allows users to view and analyze transactions on the Rollup chain. Limitations Although this is a bake-off, we're not going to sugar coat anything. Arbitrum Rollup, like any Optimistic Rollup protocol, does have one limitation, and that's the delay on withdrawals. As for the concrete length of the delay, we've done a good deal of internal modeling and have blogged about this as well. Our current modeling suggests a 3-hour delay is sufficient (but as discussed in the linked post there is a tradeoff space between the length of the challenge period and the size of the validators’ deposit). Note that this doesn't mean that the chain is delayed for three hours. Arbitrum Rollup supports pipelining of execution, which means that validators can keep building new states even while previous ones are “in the pipeline” for confirmation. As the challenge delays expire for each update, a new state will be confirmed (read more about this here). So activity and progress on the chain are not delayed by the challenge period. The only thing that's delayed is the consummation of withdrawals. Recall though that any single honest validator knows immediately (at the speed of L1 finality) which state updates are correct and can guarantee that they will eventually be confirmed, so once a valid withdrawal has been requested on-chain, every honest party knows that the withdrawal will definitely happen. There's a natural place here for a liquidity market in which a validator (or someone who trusts a validator) can provide withdrawal loans for a small interest fee. This is a no-risk business for them as they know which withdrawals will be confirmed (and can force their confirmation trustlessly no matter what anyone else does) but are just waiting for on-chain finality. 3. The recipe: How Arbitrum Rollup works For a description of the technical components of Arbitrum Rollup and how they interact to create a highly scalable protocol with a developer experience that is identical to Ethereum, please refer to the following documents: Arbitrum Rollup Whitepaper Arbitrum academic paper (describes a previous version of Arbitrum) 4. Developer docs and APIs For full details about how to set up and interact with an Arbitrum Rollup chain or validator, please refer to our developer docs, which can be found at https://developer.offchainlabs.com/. Note that the Arbitrum version described on that site is older and will soon be replaced by the version we are entering in Reddit Bake-Off, which is still undergoing internal testing before public release. 5. Who are the validators? As with any Layer 2 protocol, advancing the protocol correctly requires at least one validator (sometimes called block producers) that is honest and available. A natural question is: who are the validators? Recall that the validator set for an Arbitrum chain is open and permissionless; anyone can start or stop validating at will. (A useful analogy is to full nodes on an L1 chain.) But we understand that even though anyone can participate, Reddit may want to guarantee that highly reputable nodes are validating their chain. Reddit may choose to validate the chain themselves and/or hire third-party validators.To this end, we have begun building a marketplace for validator-for-hire services so that dapp developers can outsource validation services to reputable nodes with high up-time. We've announced a partnership in which Chainlink nodes will provide Arbitrum validation services, and we expect to announce more partnerships shortly with other blockchain infrastructure providers. Although there is no requirement that validators are paid, Arbitrum’s economic model tracks validators’ costs (e.g. amount of computation and storage) and can charge small fees on user transactions, using a gas-type system, to cover those costs. Alternatively, a single party such as Reddit can agree to cover the costs of invited validators. 6. Reddit Contract Support Since Arbitrum contracts and transactions are byte-for-byte compatible with Ethereum, supporting the Reddit contracts is as simple as launching them on an Arbitrum chain. Minting. Arbitrum Rollup supports hybrid L1/L2 tokens which can be minted in L2 and then withdrawn onto the L1. An L1 contract at address A can make a special call to the EthBridge which deploys a "buddy contract" to the same address A on an Arbitrum chain. Since it's deployed at the same address, users can know that the L2 contract is the authorized "buddy" of the L1 contract on the Arbitrum chain. For minting, the L1 contract is a standard ERC-20 contract which mints and burns tokens when requested by the L2 contract. It is paired with an ERC-20 contract in L2 which mints tokens based on whatever programmer provided minting facility is desired and burns tokens when they are withdrawn from the rollup chain. Given this base infrastructure, Arbitrum can support any smart contract based method for minting tokens in L2, and indeed we directly support Reddit's signature/claim based minting in L2. Batch minting. What's better than a mint cookie? A whole batch! In addition to supporting Reddit’s current minting/claiming scheme, we built a second minting design, which we believe outperforms the signature/claim system in many scenarios. In the current system, Reddit periodically issues signed statements to users, who then take those statements to the blockchain to claim their tokens. An alternative approach would have Reddit directly submit the list of users/amounts to the blockchain and distribute the tokens to the users without the signature/claim process. To optimize the cost efficiency of this approach, we designed an application-specific compression scheme to minimize the size of the batch distribution list. We analyzed the data from Reddit's previous distributions and found that the data is highly compressible since token amounts are small and repeated, and addresses appear multiple times. Our function groups transactions by size, and replaces previously-seen addresses with a shorter index value. We wrote client code to compress the data, wrote a Solidity decompressing function, and integrated that function into Reddit’s contract running on Arbitrum. When we ran the compression function on the previous Reddit distribution data, we found that we could compress batched minting data down to to 11.8 bytes per minting event (averaged over a 6-month trace of Reddit’s historical token grants)compared with roughly 174 bytes of on-chain data needed for the signature claim approach to minting (roughly 43 for an RLP-encoded null transaction + 65 for Reddit's signature + 65 for the user's signature + roughly 8 for the number of Points) . The relative benefit of the two approaches with respect to on-chain call data cost depends on the percentage of users that will actually claim their tokens on chain. With the above figures, batch minting will be cheaper if roughly 5% of users redeem their claims. We stress that our compression scheme is not Arbitrum-specific and would be beneficial in any general-purpose smart contract platform. 8. Benchmarks and costs In this section, we give the full costs of operating the Reddit contracts on an Arbitrum Rollup chain including the L1 gas costs for the Rollup chain, the costs of computation and storage for the L2 validators as well as the capital lockup requirements for staking. Arbitrum Rollup is still on testnet, so we did not run mainnet benchmarks. Instead, we measured the L1 gas cost and L2 workload for Reddit operations on Arbitrum and calculated the total cost assuming current Ethereum gas prices. As noted below in detail, our measurements do not assume that Arbitrum is consuming the entire capacity of Ethereum. We will present the details of our model now, but for full transparency you can also play around with it yourself and adjust the parameters, by copying the spreadsheet found here. Our cost model is based on measurements of Reddit’s contracts, running unmodified (except for the addition of a batch minting function) on Arbitrum Rollup on top of Ethereum. On the distribution of transactions and frequency of assertions. Reddit's instructions specify the following minimum parameters that submissions should support: Over a 5 day period, your scaling PoC should be able to handle:
100,000 point claims (minting & distributing points)
75,000 one-off points burning
We provide the full costs of operating an Arbitrum Rollup chain with this usage under the assumption that tokens are minted or granted to users in batches, but other transactions are uniformly distributed over the 5 day period. Unlike some other submissions, we do not make unrealistic assumptions that all operations can be submitted in enormous batches. We assume that batch minting is done in batches that use only a few percent on an L1 block’s gas, and that other operations come in evenly over time and are submitted in batches, with one batch every five minutes to keep latency reasonable. (Users are probably already waiting for L1 finality, which takes at least that long to achieve.) We note that assuming that there are only 300,000 transactions that arrive uniformly over the 5 day period will make our benchmark numbers lower, but we believe that this will reflect the true cost of running the system. To see why, say that batches are submitted every five minutes (20 L1 blocks) and there's a fixed overhead of c bytes of calldata per batch, the cost of which will get amortized over all transactions executed in that batch. Assume that each individual transaction adds a marginal cost of t. Lastly assume the capacity of the scaling system is high enough that it can support all of Reddit's 300,000 transactions within a single 20-block batch (i.e. that there is more than c + 300,000*t byes of calldata available in 20 blocks). Consider what happens if c, the per-batch overhead, is large (which it is in some systems, but not in Arbitrum). In the scenario that transactions actually arrive at the system's capacity and each batch is full, then c gets amortized over 300,000 transactions. But if we assume that the system is not running at capacity--and only receives 300,000 transactions arriving uniformly over 5 days-- then each 20-block assertion will contain about 200 transactions, and thus each transaction will pay a nontrivial cost due to c. We are aware that other proposals presented scaling numbers assuming that 300,000 transactions arrived at maximum capacity and was executed in a single mega-transaction, but according to our estimates, for at least one such report, this led to a reported gas price that was 2-3 orders of magnitude lower than it would have been assuming uniform arrival. We make more realistic batching assumptions, and we believe Arbitrum compares well when batch sizes are realistic. Our model. Our cost model includes several sources of cost:
L1 gas costs: This is the cost of posting transactions as calldata on the L1 chain, as well as the overhead associated with each batch of transactions, and the L1 cost of settling transactions in the Arbitrum protocol.
Validator’s staking costs: In normal operation, one validator will need to be staked. The stake is assumed to be 0.2% of the total value of the chain (which is assumed to be $1 per user who is eligible to claim points). The cost of staking is the interest that could be earned on the money if it were not staked.
Validator computation and storage: Every validator must do computation to track the chain’s processing of transactions, and must maintain storage to keep track of the contracts’ EVM storage. The cost of computation and storage are estimated based on measurements, with the dollar cost of resources based on Amazon Web Services pricing.
It’s clear from our modeling that the predominant cost is for L1 calldata. This will probably be true for any plausible rollup-based system. Our model also shows that Arbitrum can scale to workloads much larger than Reddit’s nominal workload, without exhausting L1 or L2 resources. The scaling bottleneck will ultimately be calldata on the L1 chain. We believe that cost could be reduced substantially if necessary by clever encoding of data. (In our design any compression / decompression of L2 transaction calldata would be done by client software and L2 programs, never by an L1 contract.) 9. Status of Arbitrum Rollup Arbitrum Rollup is live on Ethereum testnet. All of the code written to date including everything included in the Reddit demo is open source and permissively licensed under the Apache V2 license. The first testnet version of Arbitrum Rollup was released on testnet in February. Our current internal version, which we used to benchmark the Reddit contracts, will be released soon and will be a major upgrade. Both the Arbitrum design as well as the implementation are heavily audited by independent third parties. The Arbitrum academic paper was published at USENIX Security, a top-tier peer-reviewed academic venue. For the Arbitrum software, we have engaged Trail of Bits for a security audit, which is currently ongoing, and we are committed to have a clean report before launching on Ethereum mainnet. 10. Reddit Universe Arbitrum Rollup Chain The benchmarks described in this document were all measured using the latest internal build of our software. When we release the new software upgrade publicly we will launch a Reddit Universe Arbitrum Rollup chain as a public demo, which will contain the Reddit contracts as well as a Uniswap instance and a Connext Hub, demonstrating how Community Points can be integrated into third party apps. We will also allow members of the public to dynamically launch ecosystem contracts. We at Offchain Labs will cover the validating costs for the Reddit Universe public demo. If the folks at Reddit would like to evaluate our software prior to our public demo, please email us at [email protected] and we'd be more than happy to provide early access. 11. Even more scaling: Arbitrum Sidechains Rollups are an excellent approach to scaling, and we are excited about Arbitrum Rollup which far surpasses Reddit's scaling needs. But looking forward to Reddit's eventual goal of supporting hundreds of millions of users, there will likely come a time when Reddit needs more scaling than any Rollup protocol can provide. While Rollups greatly reduce costs, they don't break the linear barrier. That is, all transactions have an on-chain footprint (because all calldata must be posted on-chain), albeit a far smaller one than on native Ethereum, and the L1 limitations end up being the bottleneck for capacity and cost. Since Ethereum has limited capacity, this linear use of on-chain resources means that costs will eventually increase superlinearly with traffic. The good news is that we at Offchain Labs have a solution in our roadmap that can satisfy this extreme-scaling setting as well: Arbitrum AnyTrust Sidechains. Arbitrum Sidechains are similar to Arbitrum Rollup, but deviate in that they name a permissioned set of validators. When a chain’s validators agree off-chain, they can greatly reduce the on-chain footprint of the protocol and require almost no data to be put on-chain. When validators can't reach unanimous agreement off-chain, the protocol reverts to Arbitrum Rollup. Technically, Arbitrum Sidechains can be viewed as a hybrid between state channels and Rollup, switching back and forth as necessary, and combining the performance and cost that state channels can achieve in the optimistic case, with the robustness of Rollup in other cases. The core technical challenge is how to switch seamlessly between modes and how to guarantee that security is maintained throughout. Arbitrum Sidechains break through this linear barrier, while still maintaining a high level of security and decentralization. Arbitrum Sidechains provide the AnyTrust guarantee, which says that as long as any one validator is honest and available (even if you don't know which one will be), the L2 chain is guaranteed to execute correctly according to its code and guaranteed to make progress. Unlike in a state channel, offchain progress does not require unanimous consent, and liveness is preserved as long as there is a single honest validator. Note that the trust model for Arbitrum Sidechains is much stronger than for typical BFT-style chains which introduce a consensus "voting" protocols among a small permissioned group of validators. BFT-based protocols require a supermajority (more than 2/3) of validators to agree. In Arbitrum Sidechains, by contrast, all you need is a single honest validator to achieve guaranteed correctness and progress. Notice that in Arbitrum adding validators strictly increases security since the AnyTrust guarantee provides correctness as long as any one validator is honest and available. By contrast, in BFT-style protocols, adding nodes can be dangerous as a coalition of dishonest nodes can break the protocol. Like Arbitrum Rollup, the developer and user experiences for Arbitrum Sidechains will be identical to that of Ethereum. Reddit would be able to choose a large and diverse set of validators, and all that they would need to guarantee to break through the scaling barrier is that a single one of them will remain honest. We hope to have Arbitrum Sidechains in production in early 2021, and thus when Reddit reaches the scale that surpasses the capacity of Rollups, Arbitrum Sidechains will be waiting and ready to help. While the idea to switch between channels and Rollup to get the best of both worlds is conceptually simple, getting the details right and making sure that the switch does not introduce any attack vectors is highly non-trivial and has been the subject of years of our research (indeed, we were working on this design for years before the term Rollup was even coined). 12. How Arbitrum compares We include a comparison to several other categories as well as specific projects when appropriate. and explain why we believe that Arbitrum is best suited for Reddit's purposes. We focus our attention on other Ethereum projects. Payment only Rollups. Compared to Arbitrum Rollup, ZK-Rollups and other Rollups that only support token transfers have several disadvantages:
As outlined throughout the proposal, we believe that the entire draw of Ethereum is in its rich smart contracts support which is simply not achievable with today's zero-knowledge proof technology. Indeed, scaling with a ZK-Rollup will add friction to the deployment of smart contracts that interact with Community Points as users will have to withdraw their coins from the ZK-Rollup and transfer them to a smart contract system (like Arbitrum). The community will be best served if Reddit builds on a platform that has built-in, frictionless smart-contract support.
All other Rollup protocols of which we are aware employ a centralized operator. While it's true that users retain custody of their coins, the centralized operator can often profit from censoring, reordering, or delaying transactions. A common misconception is that since they're non-custodial protocols, a centralized sequencer does not pose a risk but this is incorrect as the sequencer can wreak havoc or shake down users for side payments without directly stealing funds.
Sidechain type protocols can eliminate some of these issues, but they are not trustless. Instead, they require trust in some quorum of a committee, often requiring two-third of the committee to be honest, compared to rollup protocols like Arbitrum that require only a single honest party. In addition, not all sidechain type protocols have committees that are diverse, or even non-centralized, in practice.
Plasma-style protocols have a centralized operator and do not support general smart contracts.
13. Concluding Remarks While it's ultimately up to the judges’ palate, we believe that Arbitrum Rollup is the bakeoff choice that Reddit kneads. We far surpass Reddit's specified workload requirement at present, have much room to optimize Arbitrum Rollup in the near term, and have a clear path to get Reddit to hundreds of millions of users. Furthermore, we are the only project that gives developers and users the identical interface as the Ethereum blockchain and is fully interoperable and tooling-compatible, and we do this all without any new trust assumptions or centralized components. But no matter how the cookie crumbles, we're glad to have participated in this bake-off and we thank you for your consideration. About Offchain Labs Offchain Labs, Inc. is a venture-funded New York company that spun out of Princeton University research, and is building the Arbitrum platform to usher in the next generation of scalable, interoperable, and compatible smart contracts. Offchain Labs is backed by Pantera Capital, Compound VC, Coinbase Ventures, and others. Leadership Team Ed Felten Ed Felten is Co-founder and Chief Scientist at Offchain Labs. He is on leave from Princeton University, where he is the Robert E. Kahn Professor of Computer Science and Public Affairs. From 2015 to 2017 he served at the White House as Deputy United States Chief Technology Officer and senior advisor to the President. He is an ACM Fellow and member of the National Academy of Engineering. Outside of work, he is an avid runner, cook, and L.A. Dodgers fan. Steven Goldfeder Steven Goldfeder is Co-founder and Chief Executive Officer at Offchain Labs. He holds a PhD from Princeton University, where he worked at the intersection of cryptography and cryptocurrencies including threshold cryptography, zero-knowledge proof systems, and post-quantum signatures. He is a co-author of Bitcoin and Cryptocurrency Technologies, the leading textbook on cryptocurrencies, and he has previously worked at Google and Microsoft Research, where he co-invented the Picnic signature algorithm. When not working, you can find Steven spending time with his family, taking a nature walk, or twisting balloons. Harry Kalodner Harry Kalodner is Co-founder and Chief Technology Officer at Offchain Labs where he leads the engineering team. Before the company he attended Princeton as a Ph.D candidate where his research explored economics, anonymity, and incentive compatibility of cryptocurrencies, and he also has worked at Apple. When not up at 3:00am writing code, Harry occasionally sleeps.
You have some bitcoins in your wallet and want to spend them on your daily purchases. But what would that look like in a world where Visa, Mastercard and other financial services still dominate the market? The ability for bitcoin to compete with other payment systems has long been up for debate in the cryptocurrency community. When Satoshi Nakamoto programmed the blocks to have a size limit of approximately 1MB each to prevent network spam, he also created the problem of bitcoin illiquidity. Since each block takes an average of 10 minutes to process, only a small number of transactions can go through at a time. For a system that many claimed could replace fiat payments, this was a big barrier. While Visa handles around 1,700 transactions a second, bitcoin could process up to 7. An increase in demand would inevitably lead to an increase in fees, and bitcoin’s utility would be limited even further. The scaling debate has unleashed a wave of technological innovation in the search of workarounds. While significant progress has been made, a sustainable solution is still far from clear. A simple solution initially appeared to be an increase in the block size. Yet that idea turned out to be not simple at all. First, there was no clear agreement as to how much it should be increased by. Some proposals advocated for 2MB, another for 8MB, and one wanted to go as high as 32MB. The core development team argued that increasing the block size at all would weaken the protocol’s decentralization by giving more power to miners with bigger blocks. Plus, the race for faster machines could eventually make bitcoin mining unprofitable. Also, the number of nodes able to run a much heavier blockchain could decrease, further centralizing a network that depends on decentralization. Second, not everyone agrees on this method of change. How do you execute a system-wide upgrade when participation is decentralized? Should everyone have to update their bitcoin software? What if some miners, nodes and merchants don’t? And finally, bitcoin is bitcoin, why mess with it? If someone didn’t like it, they were welcome to modify the open-source code and launch their own coin. One of the earliest solutions to this issue was proposed by developer Pieter Wiulle in 2015. It’s called Segregated Witness, or SegWit. This process would increase the capacity of the bitcoin blocks without changing their size limit, by altering how the transaction data was stored. SegWit was deployed on the bitcoin network in August 2017 via a soft fork to make it compatible with nodes that did not upgrade. While many wallets and other bitcoin services are gradually adjusting their software, others are reluctant to do so because of the perceived risk and cost. Several industry players argued that SegWit didn’t go far enough – it might help in the short term, but sooner or later bitcoin would again be up against a limit to its growth. In 2017, coinciding with CoinDesk’s Consensus conference in New York, a new approach was revealed: Segwit2X. This idea – backed by several of the sector’s largest exchanges – combined SegWit with an increase in the block size to 2MB, effectively multiplying the pre-SegWit transaction capacity by a factor of 8. Far from solving the problem, the proposal created a further wave of discord. The manner of its unveiling (through a public announcement rather than an upgrade proposal) and its lack of replay protection (transactions could happen on both versions, potentially leading to double spending) rankled many. And the perceived redistribution of power away from developers towards miners and businesses threatened to cause a fundamental split in the community. Other technological approaches are being developed as a potential way to increase capacity. Schnorr signatures offer a way to consolidate signature data, reducing the space it takes up within a bitcoin block (and enhancing privacy). Combined with SegWit, this could allow a much greater number of transactions, without changing the block size limit And work is proceeding on the lightning network, a second layer protocol that runs on top of bitcoin, opening up channels of fast microtransactions that only settle on the bitcoin network when the channel participants are ready. Adoption of the SegWit upgrade is slowly spreading throughout the network, increasing transaction capacity and lowering fees. Progress is accelerating on more advanced solutions such as lightning, with transactions being sent on testnets (as well as some using real bitcoin). And the potential of Schnorr signatures is attracting increasing attention, with several proposals working on detailing functionality and integration. While bitcoin’s use as a payment mechanism seems to have taken a back seat to its value as an investment asset, the need for a greater number of transactions is still pressing as the fees charged by the miners for processing are now more expensive than fiat equivalents. More importantly, the development of new features that enhance functionality is crucial to unlocking the potential of the underlying blockchain technology.
AMA Recap telos Foundation with Crypto Hunters On August 02, 2020 at 12:00 WIB Indonesia Time / August 01 2020 at 10:00 PM ( PST ) in the Crypto Hunter Telegram Group, AMA TELOS started with Mr.Douglas as guest speaker and Gus Fahlev from Crypto Hunters as moderator. When campaigning, 10 lucky AMA participants when asking questions on Google forms and AMA sessions will get a total TELOS ( TLOS ) prize of $100. The following is a summary of AMA questions and answers announced by the moderator and Segment 1 Question 1: Can you explain us, what is Telos? Answer: Telos is a blockchain platform for smart contracts. It is a low latencya new block every half second, high capacitycurrently in the top 2 blockchains in transactions per day, according to Blocktivity.info, and no transaction fee blockchain. Telos also has many unique features that allow developers to make better, dapps, such as our Telos Decide governance engine. Question 2: what ecosystem is used by telos? Answer: Telos is its own Layer-1 blockchain, not a token on another blockchain. The technology behind Telos is EOSIO, the same technology used by EOS and WAX, for example. Question 3: I see that Telos uses EOSIO platform, what are the very significant advantages that distinguish Telos from other projects? Answer: Telos uses the EOSIO platform but we have built several additional tools. Some of these add more security and resiliency to the blockchain, such as testing block producers and removing non-performant ones, but most are related to development. Telos provides attractive development tools that arent available elsewhere. Telos Decide is a governance platform that lets any group create self-governance tools easily. These run on Telos at very little cost and can provide all kinds of voting, elections, initiative ballots, committee management and funds allocation. Telos also has Telos EVM, an Ethereum virtual machine that can run Ethereum Solidity contracts at hundreds of times the speed of Ethereum and with no costs. Another Telos technology that is deploying soon is dStor, which is a decentralized cloud storage system associated with Telos so that dapps can store files controlled by blockchain contracts. Question 4: At what stage is Teloa Road Map now? what are the latest updates currently being realized? Answer: Telos launched its mainnet in December 2018 and has so far produced over 100,000,000 blocks without ever stopping or rolling back the chain. This is likely a record for a public blockchain. We have an ongoing group Telos Core Developers who build and maintain the code and are paid by our Telos Works funding system that is voted by the Telos token holders. Telos is a leader in blockchain governance and regularly amends its governance rules based on smart contract powered voting called Telos Amend. You can see the current Telos governance rules as stored live on the blockchain at tbnoa.org. The most recent updates were adding new features to Telos Decide to make it more powerful, implementing EOSIO v2.0 which increased the capacity of Telos about 8-10 times what it previously was, and implementing Telos EVM on our Testnet. We are currently working on better interfaces for Telos Decide voting, and building more infrastructure around Telos EVM so that it is ready to deploy on our mainnet. Question 5: Is telos currently available on an exchange? and is it ready to be traded? Answer: Telos has been trading on exchanges for over a year. The largest exchanges are Probit, CoinTiger, CoinLim, and P2PB2B. Other exchanges include Newdex and Alcor. We expect to be listed on larger exchanges in the near future. Question 6: Now is the time when defi tokens begin to develop, can telos be categorized as a defi project? and what strategies for this year and in the years to come prepared by telos? Answer: Telos is a smart contract platform, but it already has many DeFi tools built for it including REX staking rewards with a current yield of ~19% APR, smart contract controlled token swaps (like Bancor) with no counterparty called Telos Swaps, a common liquidity pool/order book shared by multiple DEXs to improve liquidity called EvolutionDEX. Wrapped BTC, ETH, XRP, EOS, and other tokens can be brought to Telos and exchanged or used via smart contracts through Transledger. We have more DeFi tools coming all the time including two new offerings in the next few weeks that will be the first of their kind. Question 7: Governance is an important topic in blockchain and Telos is considered a leader in this area. Why is that? Answer: Telos is among the top blockchain projects in terms of how it empowers its users to guide the growth of the chainalong the likes of Tezos or new DeFi tokens that offer governance coins. Telos users continuously elect the validating nodes, called Block Producers, that operate the network based on a set of governance documents such as the Telos Blockchain Network Operating Agreement (TBNOA). These are all stored entirely on-chain (viewable at tbnoa.org) and can be modified by smart contract through blockchain voting using Telos Amend. You can see examples of this at https://chainspector.io/governance/ratify-proposals Telos also has a robust user-voted funding mechanism called Telos Works that has funded many projects and is one of the more successful blockchain incubators around. Voting for all of these can be done in a number of ways including block explorers, wallets like Sqrl (desktop) and Telos Wallet (mobile), telos.net and Chainspector (https://chainspector.io/governance/telos-works). But Telos goes beyond any other chain-level governance by making all of these features and more available to any dapp on Telos through Telos Decide governance engine, making it easy for any dapp or DAO to add robust, highly customized voting. Segment 2 from google form Question 1: Defi projects are now trending whether telos will also go to Defi projects, to increase investors or the community? Answer: Yes, we have several DeFi tools on Telos that can work together: Telos Swaps is an automated, zero-counterparty token swapping smart contract where you can exchange any Telos tokens you may want at any time. Telos has DEXs and uses a common order book called EvolutionDEX that's available to any DEX so that a buy order on one can be matched against a sell order on another. This greatly increases liquidity for traders. We have staking rewards though the Resource EXchange (REX) with rewards currently at about 19% APR. We also have "wrapped" BTC, ETH, and other tokens that can be traded on Telos or used by its smart contracts at half-second transaction times with no transaction fees. This makes Telos a Bitcoin or Ethereum second layer or state channel that's much faster even than Lightning Network and has no fees once the BTC has been brought to Telos. Question 2: Telos aim is to build a new global economy could you explain how whole ecosystem works? There are already many centralized competitors so what is decentralization aspect in telos? Answer: Telos is one of the most decentralized blockchain's in the world. It is operated by 51 validators (block producers) who validate blocks in any month. These are voted for on an ongoing basis by Telos account holders. Telos is also economically decentralized with no large whales like Bitcoin, Ethereum, XRP or EOS because Telos never performed an ICO and limited the size of genesis accounts to 40,000 TLOS max. Telos is also geographically decentralized with users and block producers on every continent but Antarctica and in numerous countries. The is a large amount in North America and Western Europe, but also in Asia, Australia, and large contingents in Latin America and Africa. Telos has had a Block Producer in Indonesia since the beginning and some dapps on Telos are based in Indonesia as well, like SEEDS, for example. Question 3: Most investors focus only on the token price in the short term instead of the real value of the project. Can #TELOStell me the benefits for investors holding #TELOSthe long term? Answer: That's true about crypto speculators and traders, certainly. Traders are usually looking for coins with good positive momentuum that they hope will continue. But these are often pump and dumps where a few people get in early, pump the price, and then get out at the expense of new investors. That's very unfortunate. Telos isn't like this. One reason is that there aren't large whales who can easily manipulate the price. Telos seems to be greatly undervalued compared to its peers. Telos has capacity like EOS and well above XRP, XML, Tron, Ethereum. But its value is miniscule relative to these. Telos is a leader in blockchain governance like Tezos but its marketcap is tiny in comparison. Telos onboarded 100,000 new accounts last month and is appearing in the leading crypto press every week with new dapps or developments. So there's some disconnect between the value of Telos and the price. In my experience, these tend to equalize once more people learn about a project. Question 4: Eos Problems and How Telos Will Solve Them? Answer: Telos originally set out to solve problems with EOS. It was successful in this and now Telos stands on it's own and our roadmap is more about empowering users. In short, these are some of the EOS problems we already solved: RAM speculation - Telos had a plan to reduce RAM speculation through a published guidance price that has been extremely successful. The RAM price is guided by market forces but has remained within 10% of the guidance price since launch. CPU resources - Telos implemented the Telos Resource Improved Management Plan many months ago which was a 7-point approach to making EIDOS-type resource mining unprofitable on Telos. It has largely been successful and Telos has not experienced any resource shortages. Exchange Collusion/Voting - Telos governance does not permit Exchanges to vote with user tokens. This prevent voting situations seen on EOS or STEEM. Block Producer collusion - Telos has minimum requirements for block producers and do not allow anyone to own more than one block producer. Those who are found doing so (there have been about 3 cases so far) have been removed and sanctioned in accordance with the rules of the TBNOA. Question 5: What ecosystems do telos use? and why telos prefers to use EOS network over BEP2 or ERC20? what layer is used telos, can you please explain? Answer: uses the EOSIO protocol because it is the fastest and most powerful in the world and it also receives the fastest upgrades and ongoing development compared to other blockchain technologies. EOS and WAX also use the EOSIO protocol but they are completely different chains. Telos is a Layer 1 protocol, meaning that it is its own blockchain that other dapps and smart contracts deploy upon. One thing that happens when a blockchain like Telos has much, much higher speed and capacity than others like Bitcoin or Ethereum is that Telos can actually run those other blockchains better on its own platform than they can natively. For example, a number of tokens can come in to Telos as wrapped tokens. BTC, ETH, XRP are all current examples of tokens that can be on Telos as wrapped tokens. Once there, these can all be moved around with half-second transaction times and no transaction fees, so they are a better second layer for Bitcoin or Ethereum than Lightning Network or Loom. Telos can also emulate other chains, which we are doing using Telos EVM which emulates the Ethereum Virtual Machine at about 300 times faster and with no gas fees or congestion compared to Ethereum native deployment. Telos can run Ethereum (Solidity) smart contracts without any changes required. Telos EVM is already deployed on the Telos Testnet and will move to our mainnet soon. So anyone who wants to run ERC-20 tokens on Telos can do so easily and they will be faster and with much less cost than running the same contract on Ethereum. Segment 3 free asking Question: I am happy to see new things created by the Telos team. Like What concept did you build in 2020 to make Telos superior? Answer: Currently, I think Telos Decide is the most unique and powerful feature we have built. There are all kinds of organizations that need to vote. Apartment buildings, school boards, unions, tribes, youth sports leagues, city councils. Voting is hard, time consuming, and expensive for many. Telos Decide makes voting easy, convenient, and transparent. That will be a major improvement and disrupt old style voting. It also goes for buisnesses and corporate governance. Even before COVID it was important, but now people can't really gather in one place so fraud-proof voting is very important. No one has the tools that Telos has. And if they try to copy us, well, we are already way out ahead working on the next features. Question: If we look about partnerships, Telos has many partnership ! so what's the importance of that partnership for Telos? And How will you protect the value of Telos to your partners or investors ?? Answer: Many of the partnerships are dapps that have decided to deploy on Telos and receive some level of help from the TCD or Telos Foundation to do so. Once a dapp deploys on a chain, it really is like a long term partnership. Many dapps will become block producers as well and join in the governance of Telos. I suspect that in a few years, most block producers will be the large dapps on the platform with just a few remaining like my company GoodBlock. Of course, we will have our own apps out as well so I guess we'll be developers too. Telos is very fiscally responsible for investors. We spend little. There has not been any actual inflation on the chain in almost a year. (The token supply has remained unchanged at about 355M TLOS) we are actively working with dapps to bring more to Telos and exchanges and other services like fiat on- and off-ramps to increase value for users. Question: In challenging crypto market condition any project is really difficult to survive and we are witnessing that there are many platforms . What is telos project plan for surviving in this long blockchain marathon? In this plan, what motivates long term investors and believers? Answer: True. While we currently have a low token price, Telos as a DPOS chain can be maintained and grow without a massive army of miners and still maintain BFT. But the risk is really not whether Telos can continue. Already there are enough dapps that if the block producers went away somehow (not gonna happen) the dapps would just run the chain themselves. But with 100,000 new users last month and new dapps all the time, we are looking to join the top 5 dapp platforms on DappRadar soon. Survival as a project is not in question. One of the big reasons is that we never did any ICO and Telos is not a company. So regulatory risks aren't there and there's no company to go bankrupt or fail. We have already developed a bootstrapped system to pay block producers and core developers. So we aren't like a company that will run out of runway sometime. Question: Could you explain what is DSTOR? What will it contribute to your ecosystem? Answer: dStor is a decentralized cloud storage system that will have the performance of AWS or Azure with much lower costs and true decentralization. It's based on a highly modified version of IPFS that we have applied for patents for our implementation. It means that dapps will be able to store data like files, images, sound, etc. in a decentralized way. Question: Trust and security is very important in any business , what makes investors , customer and users safe secure when working with TELOS?? Answer: Telos is decentralized in a way that's more like bitcoin than other blockchains (but without the whales who can manipulate price). There was never any single company that started Telos, so there's no company whose CEO could make decisions for the network. There are numerous block producers who decide on any operational issue that isn't clearly described in the TBNOA governance documents. And to get to an action, 15 of the 21 currently active BPs need to sign a multisig transaction. So that's a high threshold. But also, the TBNOA speaks to a large number of issues and so the BPs can't just make up their own rules. Since there are really no whales, no one can vote in any kind of change or bring in their own BPs with their votes. This is also very different from other chains where there are whales. Telos is not located in any one country, so our rules can't be driven by one nation's politics. All in all, this level of decentralization sets Telos apart from almost any blockchain project in existence. People don't have to trust Telos because the system is designed to make trust unnecessary.
Author: Gamals Ahmed, CoinEx Business Ambassador https://preview.redd.it/5bqakdqgl3g51.jpg?width=865&format=pjpg&auto=webp&s=b709794863977eb6554e3919b9e00ca750e3e704 A decentralized storage network that transforms cloud storage into an account market. Miners obtain the integrity of the original protocol by providing data storage and / or retrieval. On the contrary, customers pay miners to store or distribute data and retrieve it. Filecoin announced, that there will be more delays before its main network is officially launched. Filecoin developers postponed the release date of their main network to late July to late August 2020. As mentioned in a recent announcement, the Filecoin team said that the initiative completed the first round of the internal protocol security audit. Platform developers claim that the results of the review showed that they need to make several changes to the protocol’s code base before performing the second stage of the software testing process. Created by Protocol Labs, Filecoin was developed using File System (IPFS), which is a peer-to-peer data storage network. Filecoin will allow users to trade storage space in an open and decentralized market. Filecoin developers implemented one of the largest cryptocurrency sales in 2017. They have privately obtained over $ 200 million from professional or accredited investors, including many institutional investors. The main network was slated to launch last month, but in February 2020, the Philly Queen development team delayed the release of the main network between July 15 and July 17, 2020. They claimed that the outbreak of the Coronavirus (COVID-19) in China was the main cause of the delay. The developers now say that they need more time to solve the problems found during a recent codecase audit. The Filecoin team noted the following: “We have drafted a number of protocol changes to ensure that building our major network launch is safe and economically sound.” The project developers will add them to two different implementations of Filecoin (Lotus and go-filecoin) in the coming weeks. Filecoin developers conducted a survey to allow platform community members to cast their votes on three different launch dates for Testnet Phase 2 and mainnet. The team reported that the community gave their votes. Based on the vote results, the Filecoin team announced a “conservative” estimate that the second phase of the network test should begin by May 11, 2020. The main Filecoin network may be launched sometime between July 20 and August 21, 2020. The updates to the project can be found on the Filecoin Road Map. Filecoin developers stated: “This option will make us get the most important protocol changes first, and then implement the rest as protocol updates during testnet.” Filecoin is back down from the final test stage. Another filecoin decentralized storage network provider launched its catalytic test network, the final stage of the storage network test that supports the blockchain. In a blog post on her website, Filecoin said she will postpone the last test round until August. The company also announced a calibration period from July 20 to August 3 to allow miners to test their mining settings and get an idea of how competition conditions affected their rewards. Filecoin had announced earlier last month that the catalytic testnet test would precede its flagship launch. The delay in the final test also means that the company has returned the main launch window between August 31 and September 21. Despite the lack of clear incentives for miners and multiple delays, Filecoin has succeeded in attracting huge interest, especially in China. Investors remained highly speculating on the network’s mining hardware and its premium price. Mining in Filecoin In most blockchain protocols, “miners” are network participants who do the work necessary to promote and maintain the blockchain. To provide these services, miners are compensated in the original cryptocurrency. Mining in Filecoin works completely differently — instead of contributing to computational power, miners contribute storage capacity to use for dealing with customers looking to store data. Filecoin will contain several types of miners: Storage miners responsible for storing files and data on the network. Miners retrieval, responsible for providing quick tubes for file recovery. Miners repair to be carried out. Storage miners are the heart of the network. They earn Filecoin by storing data for clients, and computerizing cipher directories to check storage over time. The probability of earning the reward reward and transaction fees is proportional to the amount of storage that the Miner contributes to the Filecoin network, not the hash power. Retriever miners are the veins of the network. They earn Filecoin by winning bids and mining fees for a specific file, which is determined by the market value of the said file size. Miners bandwidth and recovery / initial transaction response time will determine its ability to close recovery deals on the network. The maximum bandwidth of the recovery miners will determine the total amount of deals that it can enter into. In the current implementation, the focus is mostly on storage miners, who sell storage capacity for FIL.
The current system specifications recommended for running the miner are:
NVIDIA-manufactured GPU (to be expanded).
SSD drive designated as large buffer (512GB +).
Large amount of RAM for data replication account (128GB +)
Compared to the hardware requirements for running a validity checker, these standards are much higher — although they definitely deserve it. Since these will not increase in the presumed future, the money spent on Filecoin mining hardware will provide users with many years of reliable service, and they pay themselves many times. Think of investing as a small business for cloud storage. To launch a model on the current data hosting model, it will cost millions of dollars in infrastructure and logistics to get started. With Filecoin, you can do the same for a few thousand dollars. Proceed to mining Deals are the primary function of the Filecoin network, and it represents an agreement between a client and miners for a “storage” contract. Once the customer decides to have a miner to store based on the available capacity, duration and price required, he secures sufficient funds in a linked portfolio to cover the total cost of the deal. The deal is then published once the mine accepts the storage agreement. By default, all Filecoin miners are set to automatically accept any deal that meets their criteria, although this can be disabled for miners who prefer to organize their deals manually. After the deal is published, the customer prepares the data for storage and then transfers it to the miner. Upon receiving all the data, the miner fills in the data in a sector, closes it, and begins to provide proofs to the chain. Once the first confirmation is obtained, the customer can make sure the data is stored correctly, and the deal has officially started. Throughout the deal, the miner provides continuous proofs to the chain. Clients gradually pay with money they previously closed. If there is missing or late evidence, the miner is punished. More information about this can be found in the Runtime, Cut and Penalties section of this page. At Filecoin, miners earn two different types of rewards for their efforts: storage fees and reward prevention. Storage fees are the fees that customers pay regularly after reaching a deal, in exchange for storing data. This fee is automatically deposited into the withdrawal portfolio associated with miners while they continue to perform their duties over time, and is locked for a short period upon receipt. Block rewards are large sums given to miners calculated on a new block. Unlike storage fees, these rewards do not come from a linked customer; Instead, the new FIL “prints” the network as an inflationary and incentive measure for miners to develop the chain. All active miners on the network have a chance to get a block bonus, their chance to be directly proportional to the amount of storage space that is currently being contributed to the network. Duration of operation, cutting and penalties “Slashing” is a feature found in most blockchain protocols, and is used to punish miners who fail to provide reliable uptime or act maliciously against the network. In Filecoin, miners are susceptible to two different types of cut: storage error cut, unanimously reduce error. Storage Error Reduction is a term used to include a wider range of penalties, including error fees, sector penalties, and termination fees. Miners must pay these penalties if they fail to provide reliability of the sector or decide to leave the network voluntarily. An error fee is a penalty that a miner incurs for each non-working day. Sector punishment: A penalty incurred by a miner of a disrupted sector for which no error was reported before the WindowPoSt inspection. The sector will pay an error fee after the penalty of the sector once the error is discovered. Termination Fee: A penalty that a miner incurs when a sector is voluntary or involuntarily terminated and removed from the network. Cutting consensus error is the penalty that a miner incurs for committing consensus errors. This punishment applies to miners who have acted maliciously against the network consensus function. Filecoin miners Eight of the top 10 Felticoin miners are Chinese investors or companies, according to the blockchain explorer, while more companies are selling cloud mining contracts and distributed file sharing system hardware. CoinDesk’s Wolfe Chao wrote: “China’s craze for Filecoin may have been largely related to the long-standing popularity of crypto mining in the country overall, which is home to about 65% of the computing power on Bitcoin at discretion.” With Filecoin approaching the launch of the mainnet blocknet — after several delays since the $ 200 million increase in 2017 — Chinese investors are once again speculating strongly about network mining devices and their premium prices. Since Protocol Labs, the company behind Filecoin, released its “Test Incentives” program on June 9 that was scheduled to start in a week’s time, more than a dozen Chinese companies have started selling cloud mining contracts and hardware — despite important details such as economics Mining incentives on the main network are still endless. Sales volumes to date for each of these companies can range from half a million to tens of millions of dollars, according to self-reported data on these platforms that CoinDesk has watched and interviews with several mining hardware manufacturers. Filecoin’s goal is to build a distributed storage network with token rewards to spur storage hosting as a way to drive wider adoption. Protocol Labs launched a test network in December 2019. But the tokens mined in the testing environment so far are not representative of the true silicon coin that can be traded when the main network is turned on. Moreover, the mining incentive economics on testnet do not represent how final block rewards will be available on the main network. However, data from Blockecoin’s blocknetin testnet explorers show that eight out of 10 miners with the most effective mining force on testnet are currently Chinese miners. These eight miners have about 15 petabytes (PB) of effective storage mining power, accounting for more than 85% of the total test of 17.9 petable. For the context, 1 petabyte of hard disk storage = 1000 terabytes (terabytes) = 1 million gigabytes (GB). Filecoin craze in China may be closely related to the long-standing popularity of crypt mining in the country overall, which is home to about 65% of the computing power on Bitcoin by estimation. In addition, there has been a lot of hype in China about foreign exchange mining since 2018, as companies promote all types of devices when the network is still in development. “Encryption mining has always been popular in China,” said Andy Tien, co-founder of 1475, one of several mining hardware manufacturers in Philquin supported by prominent Chinese video indicators such as Fenbushi and Hashkey Capital. “Even though the Velikoyen mining process is more technologically sophisticated, the idea of mining using hard drives instead of specialized machines like Bitcoin ASIC may be a lot easier for retailers to understand,” he said. Meanwhile, according to Feixiaohao, a Chinese service comparable to CoinMarketCap, nearly 50 Chinese crypto exchanges are often somewhat unknown with some of the more well-known exchanges including Gate.io and Biki — have listed trading pairs for Filecoin currency contracts for USDT. In bitcoin mining, at the current difficulty level, one segment per second (TH / s) fragmentation rate is expected to generate around 0.000008 BTC within 24 hours. The higher the number of TH / s, the greater the number of bitcoins it should be able to produce proportionately. But in Filecoin, the efficient mining force of miners depends on the amount of data stamped on the hard drive, not the total size of the hard drive. To close data in the hard drive, the Filecoin miner still needs processing power, i.e. CPU or GPU as well as RAM. More powerful processors with improved software can confine data to the hard drive more quickly, so miners can combine more efficient mining energy faster on a given day. As of this stage, there appears to be no transparent way at the network level for retail investors to see how much of the purchased hard disk drive was purchased which actually represents an effective mining force. The U.S.-based Labs Protocol was behind Filecoin’s initial coin offer for 2017, which raised an astonishing $ 200 million. This was in addition to a $ 50 million increase in private investment supported by notable venture capital projects including Sequoia, Anderson Horowitz and Union Square Ventures. CoinDk’s parent company, CoinDk, has also invested in Protocol Labs. After rounds of delay, Protocol Protocols said in September 2019 that a testnet launch would be available around December 2019 and the main network would be rolled out in the first quarter of 2020. The test started as promised, but the main network has been delayed again and is now expected to launch in August 2020. What is Filecoin mining process? Filecoin mainly consists of three parts: the storage market (the chain), the blockecin Filecoin, and the search market (under the chain). Storage and research market in series and series respectively for security and efficiency. For users, the storage frequency is relatively low, and the security requirements are relatively high, so the storage process is placed on the chain. The retrieval frequency is much higher than the storage frequency when there is a certain amount of data. Given the performance problem in processing data on the chain, the retrieval process under the chain is performed. In order to solve the security issue of payment in the retrieval process, Filecoin adopts the micro-payment strategy. In simple terms, the process is to split the document into several copies, and every time the user gets a portion of the data, the corresponding fee is paid. Types of mines corresponding to Filecoin’s two major markets are miners and warehousers, among whom miners are primarily responsible for storing data and block packages, while miners are primarily responsible for data query. After the stable operation of the major Filecoin network in the future, the mining operator will be introduced, who is the main responsible for data maintenance. In the initial release of Filecoin, the request matching mechanism was not implemented in the storage market and retrieval market, but the takeover mechanism was adopted. The three main parts of Filecoin correspond to three processes, namely the stored procedure, retrieval process, packaging and reward process. The following figure shows the simplified process and the income of the miners: The Filecoin mining process is much more complicated, and the important factor in determining the previous mining profit is efficient storage. Effective storage is a key feature that distinguishes Filecoin from other decentralized storage projects. In Filecoin’s EC consensus, effective storage is similar to interest in PoS, which determines the likelihood that a miner will get the right to fill, that is, the proportion of miners effectively stored in the entire network is proportional to final mining revenue. It is also possible to obtain higher effective storage under the same hardware conditions by improving the mining algorithm. However, the current increase in the number of benefits that can be achieved by improving the algorithm is still unknown. It seeks to promote mining using Filecoin Discover Filecoin announced Filecoin Discover — a step to encourage miners to join the Filecoin network. According to the company, Filecoin Discover is “an ever-growing catalog of numerous petabytes of public data covering literature, science, art, and history.” Miners interested in sharing can choose which data sets they want to store, and receive that data on a drive at a cost. In exchange for storing this verified data, miners will earn additional Filecoin above the regular block rewards for storing data. Includes the current catalog of open source data sets; ENCODE, 1000 Genomes, Project Gutenberg, Berkley Self-driving data, more projects, and datasets are added every day. Ian Darrow, Head of Operations at Filecoin, commented on the announcement: “Over 2.5 quintillion bytes of data are created every day. This data includes 294 billion emails, 500 million tweets and 64 billion messages on social media. But it is also climatology reports, disease tracking maps, connected vehicle coordinates and much more. It is extremely important that we maintain data that will serve as the backbone for future research and discovery”. Miners who choose to participate in Filecoin Discover may receive hard drives pre-loaded with verified data, as well as setup and maintenance instructions, depending on the company. The Filecoin team will also host the Slack (fil-Discover-support) channel where miners can learn more. Filecoin got its fair share of obstacles along the way. Last month Filecoin announced a further delay before its main network was officially launched — after years of raising funds. In late July QEBR (OTC: QEBR) announced that it had ceded ownership of two subsidiaries in order to focus all of the company’s resources on building blockchain-based mining operations. The QEBR technology team previously announced that it has proven its system as a Filecoin node valid with CPU, GPU, bandwidth and storage compatibility that meets all IPFS guidelines. The QEBR test system is connected to the main Filecoin blockchain and the already mined filecoin coin has already been tested. “The disclosure of Sheen Boom and Jihye will allow our team to focus only on the upcoming global launch of Filecoin. QEBR branch, Shenzhen DZD Digital Technology Ltd. (“ DZD “), has a strong background in blockchain development, extraction Data, data acquisition, data processing, data technology research. We strongly believe Filecoin has the potential to be a leading blockchain-based cryptocurrency and will make every effort to make QEBR an important player when Mainecoin mainnet will be launched soon”. IPFS and Filecoin Filecoin and IPFS are complementary protocols for storing and sharing data in a decentralized network. While users are not required to use Filecoin and IPFS together, the two combined are working to resolve major failures in the current web infrastructure. IPFS It is an open source protocol that allows users to store and transmit verifiable data with each other. IPFS users insist on data on the network by installing it on their own device, to a third-party cloud service (known as Pinning Services), or through community-oriented systems where a group of individual IPFS users share resources to ensure the content stays live. The lack of an integrated catalytic mechanism is the challenge Filecoin hopes to solve by allowing users to catalyze long-term distributed storage at competitive prices through the storage contract market, while maintaining the efficiency and flexibility that the IPFS network provides. Using IPFS In IPFS, the data is hosted by the required data installation nodes. For data to persist while the user node is offline, users must either rely on their other peers to install their data voluntarily or use a central install service to store data. Peer-to-peer reliance caching data may be a good thing as one or multiple organizations share common files on an internal network, or where strong social contracts can be used to ensure continued hosting and preservation of content in the long run. Most users in an IPFS network use an installation service. Using Filecoin The last option is to install your data in a decentralized storage market, such as Filecoin. In Filecoin’s structure, customers make regular small payments to store data when a certain availability, while miners earn those payments by constantly checking the integrity of this data, storing it, and ensuring its quick recovery. This allows users to motivate Filecoin miners to ensure that their content will be live when it is needed, a distinct advantage of relying only on other network users as required using IPFS alone. Filecoin, powered by IPFS It is important to know that Filecoin is built on top of IPFS. Filecoin aims to be a very integrated and seamless storage market that takes advantage of the basic functions provided by IPFS, they are connected to each other, but can be implemented completely independently of each other. Users do not need to interact with Filecoin in order to use IPFS. Some advantages of sharing Filecoin with IPFS:
Filecoin and IPFS CIDs share hash specifications.
Use libp2p by Filecoin nodes to create secure connections with each other.
Messaging between nodes and cluster propagation is facilitated in Filecoin by libp2p pubsub.
IPLD use for blockchain data structures.
Use Graphsync to transfer data between nodes.
Of all the decentralized storage projects, Filecoin is undoubtedly the most interested, and IPFS has been running stably for two years, fully demonstrating the strength of its core protocol. Filecoin’s ability to obtain market share from traditional central storage depends on end-user experience and storage price. Currently, most Filecoin nodes are posted in the IDC room. Actual deployment and operation costs are not reduced compared to traditional central cloud storage, and the storage process is more complicated. PoRep and PoSt, which has a large number of proofs of unknown operation, are required to cause the actual storage cost to be so, in the early days of the release of Filecoin. The actual cost of storing data may be higher than the cost of central cloud storage, but the initial storage node may reduce the storage price in order to obtain block rewards, which may result in the actual storage price lower than traditional central cloud storage. In the long term, Filecoin still needs to take full advantage of its P2P storage, convert storage devices from specialization to civil use, and improve its algorithms to reduce storage costs without affecting user experience. The storage problem is an important problem to be solved in the blockchain field, so a large number of storage projects were presented at the 19th Web3 Summit. IPFS is an important part of Web3 visibility. Its development will affect the development of Web3 to some extent. Likewise, Web3 development somewhat determines the future of IPFS. Filecoin is an IPFS-based storage class project initiated by IPFS. There is no doubt that he is highly expected. Resources :
Why will the switching to TkeyNet take place this year, and not later, as planned?
Let’s look at the project history. The TKEY concept dates back to October 2017, and it was in the fourth quarter of 2017 that the distributed infrastructure concept was approved. In early 2018, the formation of the TkeyNet architecture began. To make the whole course of events clear, we highlighted the main points and commented on them: The projected development period for TkeyNet is 2.5–3 years.
This forecast was made in 2018 when the development of TkeyNet began.
The course of events that was part of our strategy
Core 1.0 launch and exchange The company planned to launch a Protocol based on Core 1.0 and conduct a subsequent listing of the asset on the exchange in late 2018-in the first half of 2019. Depending on the completion of work on Core 1.0. Why launch Core 1.0? There is a fixed practice in the market when a project starts on a ready-made blockchain, and then switches to its own, for example, EOS. This project was launched based on the Ethereum blockchain, and later the transition to its Protocol was made. Our main task was to launch a Protocol with non-standard technical solutions for the market and enter the auction to expand the project audience and obtain liquidity for the asset. With an increase in the asset price, the company would be able to increase its financial resources and reinvest them in the development of the project. Thus, the launch of a blockchain-based on Core 1.0 fully met these tasks.
In Core 1.0, new transaction models introduced and multi-blockchain support implemented. The first version of the Protocol supported the inclusion of 10 separate chains. The mechanics allowed you to change the number of parallel chains in the blockchain. To increase throughput, the team implemented PostgreSQL support, instead of the typical key-value database that is present in most cryptocurrencies.
Switching to Core 2.0 during trading and then switching to TkeyNet Next, the plan was to upgrade the network to Core 2.0 and continuously modify it. The modification means the gradual implementation of functionality and standards from TkeyNet so that it is easy to make the transition from Core 2.0 to the new TkeyNet Protocol during trading on the exchange. https://preview.redd.it/zcf5vnsgg2f51.png?width=1191&format=png&auto=webp&s=d5d5e41551ccc95f8a8a401f8fd2d081f1068939 In 2019, a Core 1.0 — based system launched. The year was simultaneously busy: the first presentation of TkeyNet at APA-2019, presence at IFC-2019, work on draft laws, and at the same time, the year was quite difficult for our company, which affected the timing shifts for products and all project plans in General. The listing did not take place. Reasons for switching to TkeyNet There is a silver lining. In the period from April to May, there was positive news from developers: work on TkeyNet will be completed much earlier than planned. By the end of June, we were preparing to launch a test network based on TkeyNet, to start the final testing of all functions.
On June 22, 2020, the core 1.0 network suspended. For more information,see the link.
Shortly, we will be able to switch to TkeyNet and list the TKEY asset to crypto exchange.
Upon completion of the launch of TkeyNet, the official date of listing of TKEY on the trading platform will publish at the link:tkeycoin.com/start/;
What is TkeyNet?
We have already talked about TkeyNet in the previous article: TkeyNet-release date, a brief analysis of the system, further plans, gave examples of how the use of technology, told what products can be created based on TkeyNet, all this covered in General terms. https://preview.redd.it/olp8lviig2f51.png?width=7418&format=png&auto=webp&s=9403b97e8bd2080fb8678530dbb418053db317c3 In this publication, we share some theses so that you will gradually develop an objective picture of the new TkeyNet system and its capabilities, which many of you will be able to apply in the future in business or everyday life. From the very beginning of development, — TkeyNet was intended to improve the existing financial system, not to replace it. From a technical point of view, the system and its functionality entirely based on blockchain technology. However, this is not a classic variation, as, for example, with bitcoin, but the new implementation of It — more secure, more suitable for global use, more perfect. In simple words, our developers took the best from Bitcoin, Ethereum, Litecoin, and other market leaders, combined their pros, eliminated their cons, and modified existing solutions on the market, resulting in new technology with new features. For the user, TkeyNet is a fast payment network that allows you to store, use, and move various assets in the payment network, such as currencies, shares, real estate, and precious metals, etc. Businesses will be able to legally conduct international transfers in seconds and significantly save on transactions. For developers and startups, this means best practices, infrastructure, liquidity, and access to ready-made solutions that can complete in their products. Among competitors, TkeyNet is much faster than its predecessors, more profitable, and cheaper in terms of transactions. For businesses and financial institutions, it is an infrastructure that will significantly improve existing financial processes, from payment routing to multi-level exchange and clearing operations. If we compare the giants of the financial industry-banks, and the new paradigm — distributed payment systems, we will notice a significant difference. The total market capitalization of cryptocurrencies estimated at ≈340 billion US dollars and the capitalization of 10 world banks is 2 trillion dollars. A significant difference, don’t you agree? http://www.outsourcingportal.eu/en/bitcoin-would-rank-as-8th-largest-bank-globally-with-169-billion-in-market-capitalization You can’t argue with the numbers, and we must understand that banks remain vital objects of the financial system. Banks help us send funds within the country and abroad, and provide a lot of services, such as loans, deposits, and a lot of other services. Anyway, using cryptocurrency, users actively exchange it for Fiat currencies to pay for any formed needs. Therefore, TkeyNet will serve as a bridge between fiat and digital currencies, providing its users with best practices and tools through which we will all have access to various digital and cash at any time and anywhere in the world. The Asian Parliamentary Assembly actively raised the issue of trust and the development of financial products in underdeveloped countries. The problem in such countries is total state control of property registers. Citizens prefer to dispose of their funds in informal settings because they do not consider state systems reliable.
According to the World Bank alone, about 1.7–1.8 billion people do not have accounts in any financial institution, and about 47% of them located in developing countries. The problem of interaction between a person and a financial institution consists of three main reasons: poverty, trust issues, and geographical difficulties. With systems such as TkeyNet, it is possible to connect people and financial institutions with a single source of trust. With the use of such systems, a person does not need anything other than access to the Internet. https://www.statista.com/chart/18497/countries-with-the-highest-share-of-adults-without-a-bank-account-in-2017/
The investments that bring us all together
On the other hand, the audience of the TKEY project is quite diverse: our investors represent a variety of professions, a variety of cities, and a variety of age groups. However, one thing, nevertheless, unites us all — this thing is an investment. And therefore, some of the users may not be interested in technical details or the difference between 1.0, 2.0, or TkeyNet. But at least the thesis, the main message, must be understood by absolutely everyone.
The more popular the company’s products are on the market, the stronger it is and the development. Due to the reliability of the company, the prices of its assets grow.
From 22 to 24 July, the test network TkeyNet was successfully launched. Our team is currently actively testing the entire network and conducting a security audit. Developers are testing the network with different scenarios: security, reliability of the full system, as well as individual modules and functions. Given the different number of similar-looking formulations, but at the same time completely different from each other, some users wondered what is the difference between such concepts: Mainnet, Testnet, and TkeyNet. Testnet should consider as a demonstration network for testing, testing concepts, new features, experiments, and debugging without the risk of losing any data. Testnet is a polygon for the development team that used to improve the system and introduce new features. Mainnet (Main Network) this is a complete product, ready to use. TkeyNet is the name of the infrastructure, the entire system that we are developing, and Testnet and Mainnet are technical concepts within this system. After testing the system is complete, TkeyNet will launch. We will issue instructions on how to upgrade to the new Protocol and new software, respectively.
Testing takes place without any excesses, and the launch of TkeyNet is just around the corner.
Thank you for being with us! Follow the project news to stay up to date. If you missed the latest news, you read the notification on the site: https://tkeycoin.com/en/news/.
Dieter Fishbein, Ecosystem Development Lead, Web3 Foundation
Dan: Hey everyone, thanks for joining us for the Polkadot Launch AMA. We have Dieter Fishbein (Head of Ecosystem Development, our business development team), Logan Saether (Technical Education), and Will Pankiewicz (Master of Validators) joining us today. We had some great questions submitted in advance, and we’ll start by answering those and learning a bit about each of our guests. After we go through the pre-submitted questions, then we’ll open up the chat to live Q&A and the hosts will answer as many questions as they can. We’ll start off with Dieter and ask him a set of some business-related questions.
Dieter could you introduce yourself, your background, and your role within the Polkadot ecosystem?
Dieter: I got my start in the space as a cryptography researcher at the University of Waterloo. This is where I first learned about Bitcoin and started following the space. I spent the next four years or so on the investment team for a large asset manager where I primarily focused on emerging markets. In 2017 I decided to take the plunge and join the space full-time. I worked at a small blockchain-focused VC fund and then joined the Polkadot team just over a year ago. My role at Polkadot is mainly focused on ensuring there is a vibrant community of projects building on our technology.
Q: Adoption of Polkadot of the important factors that all projects need to focus on to become more attractive to the industry. So, what is Polkadot's plan to gain more Adoption? [sic]
A (Dieter): Polkadot is fundamentally a developer-focused product so much of our adoption strategy is focused around making Polkadot an attractive product for developers. This has many elements. Right now the path for most developers to build on Polkadot is by creating a blockchain using the Substrate framework which they will later connect to Polkadot when parachains are enabled. This means that much of our adoption strategy comes down to making Substrate an attractive tool and framework. However, it’s not just enough to make building on Substrate attractive, we must also provide an incentive to these developers to actually connect their Substrate-based chain to Polkadot. Part of this incentive is the security that the Polkadot relay chain provides but another key incentive is becoming interoperable with a rich ecosystem of other projects that connect to Polkadot. This means that a key part of our adoption strategy is outreach focused. We go out there and try to convince the best projects in the space that building on our technology will provide them with significant value-add. This is not a purely technical argument. We provide significant support to projects building in our ecosystem through grants, technical support, incubatoaccelerator programs and other structured support programs such as the Substrate Builders Program (https://www.substrate.io/builders-program). I do think we really stand out in the significant, continued support that we provide to builders in our ecosystem. You can also take a look at the over 100 Grants that we’ve given from the Web3 Foundation: https://medium.com/web3foundation/web3-foundation-grants-program-reaches-100-projects-milestone-8fd2a775fd6b
Q: On moving forward through your roadmap, what are your most important next priorities? Does the Polkadot team have enough fundamentals (Funds, Community, etc.) to achieve those milestones?
A (Dieter): I would say the top priority by far is to ensure a smooth roll-out of key Polkadot features such as parachains, XCMP and other key parts of the protocol. Our recent Proof of Authority network launch was only just the beginning, it’s crucial that we carefully and successfully deploy features that allow builders to build meaningful technology. Second to that, we want to promote adoption by making more teams aware of Polkadot and how they can leverage it to build their product. Part of this comes down to the outreach that I discussed before but a major part of it is much more community-driven and many members of the team focus on this. We are also blessed to have an awesome community to make this process easier 🙂
Q: Where can a list of Polkadot's application-specific chains can be found?
A (Dieter): The best list right now is http://www.polkaproject.com/. This is a community-led effort and the team behind it has done a terrific job. We’re also working on providing our own resource for this and we’ll share that with the community when it’s ready.
Q: Could you explain the differences and similarities between Kusama and Polkadot?
A (Dieter): Kusama is fundamentally a less robust, faster-moving version of Polkadot with less economic backing by validators. It is less robust since we will be deploying new technology to Kusama before Polkadot so it may break more frequently. It has less economic backing than Polkadot, so a network takeover is easier on Kusama than on Polkadot, lending itself more to use cases without the need for bank-like security. In exchange for lower security and robustness, we expect the cost of a parachain lease to be lower on Kusama than Polkadot. Polkadot will always be 100% focused on security and robustness and I expect that applications that deal with high-value transactions such as those in the DeFi space will always want a Polkadot deployment, I think there will be a market for applications that are willing to trade cheap, high throughput for lower security and robustness such as those in the gaming, content distribution or social networking sectors. Check out - https://polkadot.network/kusama-polkadot-comparing-the-cousins/ for more detailed info!
Q: and for what reasons would a developer choose one over the other?
A (Dieter): Firstly, I see some earlier stage teams who are still iterating on their technology choosing to deploy to Kusama exclusively because of its lower-stakes, faster moving environment where it will be easier for them to iterate on their technology and build their user base. These will likely encompass the above sectors I identified earlier. To these teams, Polkadot becomes an eventual upgrade path for them if, and when, they are able to perfect their product, build a larger community of users and start to need the increased stability and security that Polkadot will provide. Secondly, I suspect many teams who have their main deployment on Polkadot will also have an additional deployment on Kusama to allow them to test new features, either their tech or changes to the network, before these are deployed to Polkadot mainnet.
Logan Saether, Technical Education, Web3 Foundation
Q: Sweet, let's move over to Logan. Logan - could you introduce yourself, your background, and your role within the Polkadot ecosystem?
A (Logan): My initial involvement in the industry was as a smart contract engineer. During this time I worked on a few projects, including a reboot of the Ethereum Alarm Clock project originally by Piper Merriam. However, I had some frustrations at the time with the limitations of the EVM environment and began to look at other tools which could help me build the projects that I envisioned. This led to me looking at Substrate and completing a bounty for Web3 Foundation, after which I applied and joined the Technical Education team. My responsibilities at the Technical Education team include maintaining the Polkadot Wiki as a source of truth on the Polkadot ecosystem, creating example applications, writing technical documentation, giving talks and workshops, as well as helping initiatives such as the Thousand Validator Programme.
Q: The first technical question submitted for you was: "When will an official Polkadot mobile wallet appear?"
A (Logan): There is already an “official” wallet from Parity Technologies called the Parity Signer. Parity Signer allows you to keep your private keys on an air-gapped mobile device and to interactively sign messages using web interfaces such as Polkadot JS Apps. If you’re looking for something that is more of an interface to the blockchain as well as a wallet, you might be interested in PolkaWallet which is a community team that is building a full mobile interface for Polkadot. For more information on Parity Signer check out the website: https://www.parity.io/signe
Q: Great thanks...our next question is: If someone already developed an application to run on Ethereum, but wants the interoperability that Polkadot will offer, are there any advantages to rebuilding with Substrate to run as a parachain on the Polkadot network instead of just keeping it on Ethereum and using the Ethereum bridge for use with Polkadot?
A (Logan): Yes, the advantage you would get from building on Substrate is more control over how your application will interact with the greater Polkadot ecosystem, as well as a larger design canvas for future iterations of your application. Using an Ethereum bridge will probably have more cross chain latency than using a Polkadot parachain directly. The reason for this is due to the nature of Ethereum’s separate consensus protocol from Polkadot. For parachains, messages can be sent to be included in the next block with guarantees that they will be delivered. On bridged chains, your application will need to go through more routes in order to execute on the desired destination. It must first route from your application on Ethereum to the Ethereum bridge parachain, and afterward dispatch the XCMP message from the Polkadot side of the parachain. In other words, an application on Ethereum would first need to cross the bridge then send a message, while an application as a parachain would only need to send the message without needing to route across an external bridge.
Q: DOT transfers won't go live until Web3 removes the Sudo module and token holders approve the proposal to unlock them. But when will staking rewards start to be distributed? Will it have to after token transfers unlock? Or will accounts be able to accumulate rewards (still locked) once the network transitions to NPoS?
A (Logan): Staking rewards will be distributed starting with the transition to NPoS. Transfers will still be locked during the beginning of this phase, but reward payments are technically different from the normal transfer mechanism. You can read more about the launch process and steps at http://polkadot.network/launch-roadmap
Q: Next question is: I'm interested in how Cumulus/parachain development is going. ETA for when we will see the first parachain registered working on Kusama or some other public testnet like Westend maybe?
A (Logan): Parachains and Cumulus is a current high priority development objective of the Parity team. There have already been PoC parachains running with Cumulus on local testnets for months. The current work now is making the availability and validity subprotocols production ready in the Polkadot client. The best way to stay up to date would be to follow the project boards on GitHub that have delineated all of the tasks that should be done. Ideally, we can start seeing parachains on Westend soon with the first real parachains being deployed on Kusama thereafter. The projects board can be viewed here: https://github.com/paritytech/polkadot/projects Dan: Also...check out Basti's tweet from yesterday on the Cumulus topic: https://twitter.com/bkchstatus/1270479898696695808?s=20
Q: In what ways does Polkadot support smart contracts?
A (Logan): The philosophy behind the Polkadot Relay Chain is to be as minimal as possible, but allow arbitrary logic at the edges in the parachains. For this reason, Polkadot does not support smart contracts natively on the Relay Chain. However, it will support smart contracts on parachains. There are already a couple major initiatives out there. One initiative is to allow EVM contracts to be deployed on parachains, this includes the Substrate EVM module, Parity’s Frontier, and projects such as Moonbeam. Another initiative is to create a completely new smart contract stack that is native to Substrate. This includes the Substrate Contracts pallet, and the ink! DSL for writing smart contracts. Learn more about Substrate's compatibility layer with Ethereum smart contracts here: https://github.com/paritytech/frontier
Will Pankiewicz, Master of Validators, Parity Technologies
Q: (Dan) Thanks for all the answers. Now we’ll start going through some staking questions with Will related to validating and nominating on Polkadot. Will - could you introduce yourself, your background, and your role within the Polkadot ecosystem?
A (Will): Sure thing. Like many others, Bitcoin drew me in back in 2013, but it wasn't until Ethereum came that I took the deep dive into working in the space full time. It was the financial infrastructure aspects of cryptocurrencies I was initially interested in, and first worked on dexes, algorithmic trading, and crypto funds. I really liked the idea of "Generalized Mining" that CoinFund came up with, and started to explore the whacky ways the crypto funds and others can both support ecosystems and be self-sustaining at the same time. This drew me to a lot of interesting experiments in what later became DeFi, as well as running validators on Proof of Stake networks. My role in the Polkadot ecosystem as “Master of Validators” is ensuring the needs of our validator community get met.
Q: Cool thanks. Our first community question was "Is it still more profitable to nominate the validators with lesser stake?"
A (Will): It depends on their commission, but generally yes it is more profitable to nominate validators with lesser stake. When validators have lesser stake, when you nominate them this makes your nomination stake a higher percentage of total stake. This means when rewards get distributed, it will be split more favorably toward you, as rewards are split by total stake percentage. Our entire rewards scheme is that every era (6 hours in Kusama, 24 hours in Polkadot), a certain amount of rewards get distributed, where that amount of rewards is dependent on the total amount of tokens staked for the entire network (50% of all tokens staked is currently optimal). These rewards from the end of an era get distributed roughly equally to all validators active in the validator set. The reward given to each validator is then split between the validators and all their nominators, determined by the total stake that each entity contributes. So if you contribute to a higher percentage of the total stake, you will earn more rewards.
Q: What does priority ranking under nominator addresses mean? For example, what does it mean that nominator A has priority 1 and nominator B has priority 6?
A (Will): Priority ranking is just the index of the nomination that gets stored on chain. It has no effect on how stake gets distributed in Phragmen or how rewards get calculated. This is only the order that the nominator chose their validators. The way that stake from a nominator gets distributed from a nominator to validators is via Phragmen, which is an algorithm that will optimally put stake behind validators so that distribution is roughly equal to those that will get in the validator set. It will try to maximize the total amount at stake in the network and maximize the stake behind minimally staked validators.
Q: On Polkadot.js, what does it mean when there are nodes waiting on Polkadot?
**A (Will):**In Polkadot there is a fixed validator set size that is determined by governance. The way validators get in the active set is by having the highest amount of total stake relative to other validators. So if the validator set size is 100, the top 100 validators by total stake will be in the validator set. Those not active in the validator set will be considered “waiting”.
Q: Another question...Is it necessary to become a waiting validator node right now?
A (Will): It's not necessary, but highly encouraged if you actively want to validate on Polkadot. The longer you are in the waiting tab, the longer you get exposure to nominators that may nominate you.
Q: Will current validators for Kusama also validate for Polkadot? How strongly should I consider their history (with Kusama) when looking to nominate a good validator for DOTs?
A (Will): A lot of Kusama validators will also be validators for Polkadot, as KSM was initially distributed to DOT holders. The early Kusama Validators will also likely be the first Polkadot validators. Being a Kusama validator should be a strong indicator for who to nominate on Polkadot, as the chaos that has ensued with Kusama has allowed validators to battle test their infrastructure. Kusama validators by now are very familiar with tooling, block explorers, terminology, common errors, log formats, upgrades, backups, and other aspects of node operation. This gives them an edge against Polkadot validators that may be new to the ecosystem. You should strongly consider well known Kusama validators when making your choices as a nominator on Polkadot.
Q: Can you go into more details about the process for becoming a DOT validator? Is it similar as the KSM 1000 validators program?
A (Will): The Process for becoming a DOT validators is first to have DOTs. You cannot be a validator without DOTs, as DOTs are used to pay transaction fees, and the minimum amount of DOTs you need is enough to create a validate transaction. After obtaining enough DOTs, you will need to set up your validator infrastructure. Ideally you should have a validator node with specs that match what we call standard hardware, as well as one or more sentry nodes to help isolate the validator node from attacks. After the infrastructure is up and running, you should have your Polkadot accounts set up right with a stash bonded to a controller account, and then submit a validate transaction, which will tell the network your nodes are ready to be a part of the network. You should then try and build a community around your validator to let others know you are trustworthy so that they will nominate you. The 1000 validators programme for Kusama is a programme that gives a certain amount of nominations from the Web3 Foundation and Parity to help bootstrap a community and reputation for validators. There may eventually be a similar type of programme for Polkadot as well. Dan: Thanks a lot for all the answers, Will. That’s the end of the pre-submitted questions and now we’ll open the chat up to live Q&A, and our three team members will get through as many of your questions as possible. We will take questions related to business development, technology, validating, and staking. For those wondering about DOT: DOT tokens do not exist yet. Allocations of Polkadot's native DOT token are technically and legally non-transferable. Hence any publicized sale of DOTs is unsanctioned by Web3 Foundation and possibly fraudulent. Any official public sale of DOTs will be announced on the Web3 Foundation website. Polkadot’s launch process started in May and full network decentralization later this year, holders of DOT allocations will determine issuance and transferability. For those who participated in previous DOT sales, you can learn how to claim your DOTs here (https://wiki.polkadot.network/docs/en/claims).
Telegram Community Follow-up Questions Addressed Below
Q: Polkadot looks good but it confuses me that there are so many other Blockchain projects. What should I pay attention in Polkadot to give it the importance it deserves? What are your planning to achieve with your project?
A (Will): Personally, what I think differentiates it is the governance process. Coordinating forkless upgrades and social coordination helps stand it apart. A (Dieter): The wiki is awesome - https://wiki.polkadot.network/
Q: Over 10,000 ETH paid as a transaction fee , what if this happens on Polkadot? Is it possible we can go through governance to return it to the owner?
Q: What is the minimum ideal amount of DOT and KSM to have if you want to become a validator and how much technical knowledge do you need aside from following the docs?
A (Will): It depends on what the other validators in the ecosystem are staking as well as the validator set size. You just need to be in the top staking amount of the validator set size. So if its 100 validators, you need to be in the top 100 validators by stake.
Q: Will Web3 nominate validators? If yes, which criteria to be elected?
Q: How did you manage to prove that the consensus protocol is safe and unbreakable mathematically?
A (Dieter): We have a research teams of over a dozen scientists with PhDs and post-docs in cryptography and distributed computing who do thorough theoretical analyses on all the protocols used in Polkadot
Q: What are the prospects for NFT?
A: Already being built 🙂
Q: What will be Polkadot next roadmap for 2020 ?
A (Dieter): Building. But seriously - we will continue to add many more features and upgrades to Polkadot as well as continue to strongly focus on adoption from other builders in the ecosystem 🙂 A (Will): https://polkadot.network/launch-roadmap/ This is the launch roadmap. Ideally adding parachains and xcmp towards the end of the year
Q: How Do you stay active in terms of marketing developments during this PANDEMIC? Because I'm sure you're very excited to promote more after this settles down.
A (Dan): The main impact of covid was the impact on in-person events. We have been very active on Crowdcast for webinars since 2019, so it was quite the smooth transition to all-online events. You can see our 40+ past event recordings and follow us on Crowdcast here: https://www.crowdcast.io/polkadot. If you're interested in following our emails for updates (including online events), subscribe here: https://info.polkadot.network/subscribe
Q: Hi, who do you think is your biggest competitor in the space?
A (Dan): Polkadot is a metaprotocol that hasn't been seen in the industry up until this point. We hope to elevate the industry by providing interoperability between all major public networks as well as private blockchains.
Q: Is Polkadot a friend or competitor of Ethereum?
A: Polkadot aims to elevate the whole blockchain space with serious advancements in interoperability, governance and beyond :)
Q: When will there be hardware wallet support?
A (Will): Parity Signer works well for now. Other hardware wallets will be added pretty soon
Q: What are the attractive feature of DOT project that can attract any new users ?
A: We are working on integrations on all the biggest and best wallet providers. ;)
Q: What event/call can we track to catch a switch to nPOS? Is it only force_new_era call? Thanks.
A (Will): If you're on riot, useful channels to follow for updates like this are #polkabot:matrix.org and #polkadot-announcements:matrix.parity.io A (Logan): Yes this is the trigger for initiating the switch to NPoS. You can also poll the ForceEra storage for when it changes to ForceNew.
Q: What strategy will the Polkadot Team use to make new users trust its platform and be part of it?
Q: What problems do you see occurring in the blockchain industry nowadays and how does your project aims to solve these problems?
A (Will): Governance I see as a huge problem. For example upgrading Bitcoin and making decisions for changing things is a very challenging process. We have robust systems of on-chain governance to help solve these coordination problems
Q: How involved are the Polkadot partners? Are they helping with the development?
Q: Can you explain the role of the treasury in Polkadot?
A (Will): The treasury is for projects or people that want to build things, but don't want to go through the formal legal process of raising funds from VCs or grants or what have you. You can get paid by the community to build projects for the community. A: There’s a whole section on the wiki about the treasury and how it functions here https://wiki.polkadot.network/docs/en/mirror-learn-treasury#docsNav
Q: Any plan to introduce Polkadot on Asia, or rising market on Asia?
**A (Will):**We're globally focused
Q: What kind of impact do you expect from the Council? Although it would be elected by token holders, what kind of people you wish to see there?
A (Will): Community focused individuals like u/jam10o that want to see cool things get built and cool communities form If you have further questions, please ask in the official Polkadot Telegram channel.
Oh, it's actually bitcoin2x testnet, not bitcoin core testnet. That's very confusing actually. Please note it in the answer – Karel Bílek Oct 5 '17 at 19:50. Oh sorry, I didn't notice that! I haven't got around using it yet so I didn't know. – Celsiuss Oct 6 '17 at 20:12. This not working now. – kodmanyagha Dec 20 '17 at 18:53. the hostname does not resolve. I tried also just 'testnet ... SoChain. Wow. The Fastest Bitcoin Testnet Block Reader. Price: 0.00 USD/฿T: Hashrate: 197.17 TH/s: Activity-.-- TX/min: Unconfirmed Txs In this post, I show you how to set up your own Bitcoin Core for Testnet, and learn how to work with your own wallet and address for Bitcoin beginners (developers). Prerequisite Settings Here I’ll run Bitcoin Core on Ubuntu 18.04 in Microsoft Azure. Testnet is an alternative Bitcoin blockchain, to be used for testing.Testnet coins are separate and distinct from actual bitcoins, and are never supposed to have any value. This allows application developers or bitcoin testers to experiment, without having to use real bitcoins or worrying about breaking the main bitcoin chain. Height Age Transactions Total Sent Total Fees Block Size (in bytes) 1864748: 2020-10-26T02:32:54.263Z: 24: 21.646 BTC: 0.001 BTC: 4,803: 1864747: 2020-10-26T02:28:42.842Z
Bitcoin Testnet Sandbox and Faucet brings all the coins to developers and interesting insights of the Testnet Blockchain. The testnet is an alternative Bitcoin block chain, to be used for testing. What isTestnet and Mainnet in Blockchain? Basic knowledge of Testnet and Mainnet Join our telegram channel for more updates: https://t.me/CryptoClubAlerts We’ll stop supporting this browser soon. For the best experience please update your browser. In this talk, Navin Kabra , Co-Founder and CTO at ReliScore.com and Creator of PuneTech.com, has given an overview of what is Bitcoin, how it works, what is the Blockchain, what is the fundamental ... How To Fund Your Blockchain Wallet With Bitcoin - Duration: 7:38. Kurt Tasche 183,815 views. 7:38. What is an API? - Duration: 3:25. MuleSoft Videos 3,099,103 views. 3:25. The Ultimate Intro to ...