Blog

Archive for the ‘information science’ category: Page 137

May 2, 2022

Codex vs Programmers: Can the Text Generator Kill Coders?

Posted by in categories: information science, robotics/AI

Codex, a deep learning model, developed by Open AI, though threatens to leave programmers jobless, it cannot be considered a complete model itself as this GPT-3 model has a long way to go.

May 1, 2022

The basics of decentralized finance

Posted by in categories: blockchains, computing, cryptocurrencies, finance, information science, mathematics

Decentralized finance is built on blockchain technology, an immutable system that organizes data into blocks that are chained together and stored in hundreds of thousands of nodes or computers belonging to other members of the network.

These nodes communicate with one another (peer-to-peer), exchanging information to ensure that they’re all up-to-date and validating transactions, usually through proof-of-work or proof-of-stake. The first term is used when a member of the network is required to solve an arbitrary mathematical puzzle to add a block to the blockchain, while proof-of-stake is when users set aside some cryptocurrency as collateral, giving them a chance to be selected at random as a validator.

To encourage people to help keep the system running, those who are selected to be validators are given cryptocurrency as a reward for verifying transactions. This process is popularly known as mining and has not only helped remove central entities like banks from the equation, but it also has allowed DeFi to open more opportunities. In traditional finance, are only offered to large organizations, for members of the network to make a profit. And by using network validators, DeFi has also been able to cut down the costs that intermediaries charge so that management fees don’t eat away a significant part of investors’ returns.

Apr 30, 2022

Deep Learning in Neuroimaging

Posted by in categories: biotech/medical, information science, robotics/AI

Our brain is constantly working to make sense of the world around us and finding patterns in it, even when we are asleep the brain is storing patterns. Making sense of the brain itself, however, has remained an intricate pursuit.

Christoff Koch, a well-known neuroscientist, famously called the human brain the “most complex object in our observable universe” [1]. Aristotle, on the other hand, thought it was the heart that gave rise to consciousness and that the brain functioned as a cooling system both practically and philosophically [2]. Theories of the brain have evolved since then, generally shaped by knowledge gathered over centuries. Historically, to analyze the brain, we had to either extract the brain from deceased people or perform invasive surgery. Progress over the past decades has led to inventions that allow us to study the brain without invasive surgeries. A few examples of imaging techniques that do not require surgery include macroscopic imaging techniques such as functional magnetic resonance imaging (fMRI) or approaches with a high temporal resolution such as electroencephalogy (EEG). Advances in treatments, such as closed-loop electrical stimulation systems, have enabled the treatment of disorders like epilepsy and more recently depression [3, 4]. Existing neuroimaging approaches can produce a considerable amount of data about a very complex organ that we still do not fully understand which has led to an interest in non-linear modeling approaches and algorithms equipped to learn meaningful features.

Continue reading “Deep Learning in Neuroimaging” »

Apr 30, 2022

Engineers use artificial intelligence to capture the complexity of breaking waves

Posted by in categories: information science, robotics/AI

Waves break once they swell to a critical height, before cresting and crashing into a spray of droplets and bubbles. These waves can be as large as a surfer’s point break and as small as a gentle ripple rolling to shore. For decades, the dynamics of how and when a wave breaks have been too complex to predict.

Now, MIT engineers have found a new way to model how waves break. The team used machine learning along with data from wave-tank experiments to tweak equations that have traditionally been used to predict wave behavior. Engineers typically rely on such equations to help them design resilient offshore platforms and structures. But until now, the equations have not been able to capture the complexity of breaking waves.

The updated model made more accurate predictions of how and when waves break, the researchers found. For instance, the model estimated a wave’s steepness just before breaking, and its energy and frequency after breaking, more accurately than the conventional wave equations.

Apr 29, 2022

Towards practical and robust DNA-based data archiving using the yin–yang codec system

Posted by in categories: chemistry, computing, information science

The yin-yang codec transcoding algorithm is proposed to improve the practicality and robustness of DNA data storage.


Given these results, YYC offers the opportunity to generate DNA sequences that are highly amenable to both the ‘writing’ (synthesis) and ‘reading’ (sequencing) processes while maintaining a relatively high information density. This is crucially important for improving the practicality and robustness of DNA data storage. The DNA Fountain and YYC algorithms are the only two known coding schemes that combine transcoding rules and screening into a single process to ensure that the generated DNA sequences meet the biochemical constraints. The comparison hereinafter thus focuses on the YYC and DNA Fountain algorithms because of the similarity in their coding strategies.

The robustness of data storage in DNA is primarily affected by errors introduced during ‘writing’ and ‘reading’. There are two main types of errors: random and systematic errors. Random errors are often introduced by synthesis or sequencing errors in a few DNA molecules and can be redressed by mutual correction using an increased sequencing depth. System atic errors refer to mutations observed in all DNA molecules, including insertions, deletions and substitutions, which are introduced during synthesis and PCR amplification (referred to as common errors), or the loss of partial DNA molecules. In contrast to substitutions (single-nucleotide variations, SNVs), insertions and deletions (indels) change the length of the DNA sequence encoding the data and thus introduce challenges regarding the decoding process. In general, it is difficult to correct systematic errors, and thus they will lead to the loss of stored binary information to varying degrees.

Continue reading “Towards practical and robust DNA-based data archiving using the yin–yang codec system” »

Apr 25, 2022

Elon Musk acquires Twitter for roughly $44 billion

Posted by in categories: cybercrime/malcode, economics, Elon Musk, information science, robotics/AI

The company’s board and the Tesla CEO hammered out the final details of his $54.20 a share bid.

The agreement marks the close of a dramatic courtship and a sharp change of heart at the social-media network.

Elon Musk acquired Twitter for $44 billion on Monday, the company announced, giving the world’s richest person command of one of its most influential social media sites — which serves as a platform for political leaders, a sounding board for experts across industries and an information hub for millions of everyday users.

Continue reading “Elon Musk acquires Twitter for roughly $44 billion” »

Apr 25, 2022

Why it’s so damn hard to make AI fair and unbiased

Posted by in categories: information science, robotics/AI

There are competing notions of fairness — and sometimes they’re incompatible, as facial recognition and lending algorithms show.

Apr 25, 2022

Quantifying Human Consciousness With the Help of AI

Posted by in categories: information science, robotics/AI

A new deep learning algorithm is able to quantify arousal and awareness in humans at the same time.

#consc… See more.


Summary: A new deep learning algorithm is able to quantify arousal and awareness in humans at the same time.

Continue reading “Quantifying Human Consciousness With the Help of AI” »

Apr 23, 2022

Growing Anomalies at the Large Hadron Collider Raise Hopes

Posted by in categories: information science, particle physics

Amid the chaotic chains of events that ensue when protons smash together at the Large Hadron Collider in Europe, one particle has popped up that appears to go to pieces in a peculiar way.

All eyes are on the B meson, a yoked pair of quark particles. Having caught whiffs of unexpected B meson behavior before, researchers with the Large Hadron Collider beauty experiment (LHCb) have spent years documenting rare collision events featuring the particles, in hopes of conclusively proving that some novel fundamental particle or effect is meddling with them.

In their latest analysis, first presented at a seminar in March, the LHCb physicists found that several measurements involving the decay of B mesons conflict slightly with the predictions of the Standard Model of particle physics — the reigning set of equations describing the subatomic world. Taken alone, each oddity looks like a statistical fluctuation, and they may all evaporate with additional data, as has happened before. But their collective drift suggests that the aberrations may be breadcrumbs leading beyond the Standard Model to a more complete theory.

Apr 22, 2022

How to generate smart games using machine learning?

Posted by in categories: information science, robotics/AI

Machine learning and machine learning algorithms are finding new applications in game building. Machine learning NPCs with machine learning processors have made it possible to have a virtual player.


Study reveals the different ways the brain parses information through interactions of waves of neural activity.