Tag Archives: Science

This is how ChatGPT works

The powerful language model ChatGPT generates texts that can hardly be distinguished from those of human authors. We explain the technology behind the hype.

Image: Programming, Free Stock Picture, MorgueFile.com.

Since the US company OpenAI released its new artificial intelligence (AI) ChatGPT for free testing at the end of November last year, users on social media have been sharing masses of examples of how the chatbot answers knowledge questions, formulates e-mails, writes poems or texts summarizes.

ChatGPT’s ability to confidently deal with natural language and to understand complex relationships with a high hit rate is seen by some observers as another milestone on the way to strong artificial intelligence – i.e. to algorithms that are on a par with human thinking ability in every respect. But how does the technology that makes all this possible work?

Six years – an AI eternity

ChatGPT is a language model, i.e. a machine learning algorithm that specializes in processing texts. ChatGPT is the latest generation in a series of language models based on the so-called Transformer model introduced in 2017. The Transformer architecture caused a stir when it was released in professional circles because it enabled specialized language models for text translation and other tasks with unprecedented power.

As early as 2018, OpenAI published the Generative Pretrained Transformer (GPT) as a modification of the Transformer with a simplified structure (PDF) . A major innovation was the idea of ​​no longer training the language model for a special task such as translation or classification of texts, for which only limited amounts of sample data are often available.

Instead, the GPT model was pre-trained on very large data sets of generic texts in order to learn statistical properties of language as such independently of the specific task. The model prepared in this way could then be effectively adapted with smaller sets of sample data for specific tasks.

The next version GPT-2 appeared in 2019 (PDF) . It was essentially a scaled-up version of the previous model with a significantly higher number of parameters and with training on correspondingly larger data sets. In contrast to the original version, GPT-2 was no longer adapted for special problems, but was able to solve many different tasks such as translating texts or answering knowledge questions simply by training with generic texts from the Internet.

With 175 billion parameters, the third generation GPT-3 (PDF) was even more extensive than GPT-2 and correspondingly more powerful. It also attracted attention beyond the AI ​​research community, particularly with its ability to write longer texts that were almost indistinguishable from those of human authors.

However, limitations of the model also became apparent, including ethical issues with objectionable or biased texts, and the habit of making grossly false statements of fact in persuasive-sounding language.

In order to remedy these shortcomings, OpenAI added a fundamentally new dimension to the training concept for its next language models InstructGPT and ChatGPT : Instead of leaving a model alone with huge amounts of text from the Internet, it was subsequently taught by human “teachers”, concrete ones To follow the instructions of the users and to make statements that are ethically justifiable and correct in terms of content. In order to ensure the effectiveness of this training, the algorithmic approach of the pure transformer model had to be expanded by a further step – the so-called reinforcement learning.

The impressive achievements of ChatGPT are the result of a whole range of different algorithms and methods as well as many tricks, some of which are very small. In this article, the focus is on providing an intuitive basic understanding of the technology without getting bogged down in too many mathematical or technical details. The links in the text refer to sources that fill in the gaps in this presentation.

Competition for SSD

Competition for SSD: 1.4 petabytes of data on future magnetic tape

Tape storage technology seemed to be falling short of expectations recently. Now it looks like it’s catching up again – thanks to a new material.

Linear Tape Open (LTO) first appeared on the market in 2000. At that time, the storage media that recorded data as encoded tracks on magnetic tape still had a capacity of 200 gigabytes.

The tapes have now reached their ninth generation and are currently able to record data volumes of 18 terabytes (uncompressed). The technology is a cost-effective way, especially for companies, to preserve critical data reliably and durably.

Image: Data Storage, Free Stock Picture, MorgueFile.com.
Image: Data Storage, Free Stock Picture, MorgueFile.com.

LTO09 still fell short of expectations

But what sounds like an incredible amount of storage space was initially a disappointment. LTO09 was originally supposed to offer 24 terabytes of space, but could not meet these expectations.

According to a new roadmap presented by the developers IBM, HPE and Quantum at the beginning of September, it should be possible to store a whopping 1,440 terabytes (1.4 petabytes) from the 14th generation, which is expected to appear in 2033/34.

Magnetic storage tapes: New material creates new possibilities

According to Heise Online , this is made possible by coating the magnetic tapes with strontium ferrite (SrFe) instead of barium ferrite (BaFe), which has been used up until now. The first prototypes have already been developed and tested by Fujifilm. The new material is to be used from LTO13.

In contrast to SSD hard drives, whose maximum capacity is currently around 100 terabytes, LTO is also significantly cheaper. While the most expensive SSD medium with a price of 40,000 US dollars causes costs of 2.5 dollars per gigabyte, LTO are only 0.01 dollars per gigabyte.

According to Sam Werner, IBM vice president of storage product management, LTO “provides organizations with a sustainable, reliable and cost-effective solution to protect and store their critical business data.”

Samsung has a new battery issue that can affect any Galaxy phone

Samsung has a new battery problem with its Galaxy smartphones. After the disaster with the Galaxy Note 7, which exploded and was recalled worldwide, older smartphones are now bloating and can become an undiscovered danger.

Image: Circuit Board Chip, Free Stock Picture, MorgueFile.com.
Image: Circuit Board Chip, Free Stock Picture, MorgueFile.com.

Samsung smartphones are bloating

Samsung has promised that after the catastrophic incidents involving the Galaxy Note 7 , the batteries will be safe and there will be no more problems. As it turns out, Samsung has an even bigger battery problem than just one model. YouTuber Mrwhosetheboss has found many Samsung phones in his smartphone collection with swollen batteries. A reaction and gas formation probably occurs inside, so that the battery swells up and the back bursts open . In this case, there is no explosion or fire. The batteries are therefore fundamentally safe.

It becomes critical if the battery is still bloating and you don’t notice it immediately. In fact, the problem tends to affect older Samsung phones that are stored with an empty battery. Our Galaxy S6 edge has also ballooned, as you can see in the cover photo above . But Mrwhosetheboss has also noticed early signs of a swelling battery on his Galaxy S20 FE and Galaxy Z Fold 2. And then it gets dangerous. If gases are produced and the battery swells, there could be a reaction and excessive heat development when charging the smartphone.

How should you store a Samsung phone?

Mrwhosetheboss gives an important tip on how to store your Samsung phones when you are no longer using them. Then you should charge the battery to about 50 percent. This should reduce the risk of the battery bloating. If you are still using an older Samsung cell phone, you should regularly check whether the battery has not already swelled up slightly. Then you shouldn’t charge your cell phone anymore and turn to Samsung. We will seek an opinion from Samsung on the matter.

Hacker Attacks on Crypto Protocols: Nearly $500M in Damage Last Quarter

Image: Hacker, Free Stock Picture, MorgueFile.com.

As the company Atlas VPN has found, hacker attacks have been particularly successful in recent months. Chainalysis had already warned of a record month.

Although the cryptocurrency sector is now much better regulated and more and more investors are taking the necessary steps to increase security – such as storing their own coins in hardware wallets – hacker attacks are still a big issue in the cryptocurrency sector, albeit only in relation to the volume traded affect a fraction of the sector.

In one quarter, hackers cause almost half a billion dollars in damage

According to Atlas VPN data , in the third quarter of 2022, criminals stole around $483 million worth of cryptocurrencies through targeted attacks. The number of hacks fell by 43 percent compared to the second quarter. In the first quarter, the damage amounted to around 1.3 billion US dollars.

Even if the damage appears large in absolute numbers and was certainly significant for those affected, in relation to the size of the crypto sector with a value of around $970 billion according to CoinMarketCap, it is not quite as dramatic as it might seem at first glance.

Ethereum, Polkadot and BNB Chain particularly affected

The hacks primarily affected the Ethereum network. A total of 11 attacks on Ethereum blockchain-based protocols caused $348 million in damage. However, considering that most protocols run on Ethereum, this is not surprising. For Polkadot, it was $52 million in just two attacks. While projects on the BNB chain have been attacked 13 times, the damage amounts to only $28 million.

It is important here that the blockchains themselves are not attacked. Instead, it is mainly smart contracts in the DeFi area that cause security gaps.

This quarter could be a record

Chainalysis also deals primarily with the damage caused by cybercrime in the cryptocurrency sector. The figures determined by Atlas VPN for the third quarter correspond to the information from Chainalysis, which expects a record month for October. As Chainalysis announced on October 12, eleven hacker attacks with damage totaling $718 million had already been registered by then.

If the trend of the month continues, the fourth quarter is likely to be the most momentous for the cryptocurrency sector. The BNB chain hack caused a stir this month , in which at least no funds were stolen from other users. Instead, the attackers created over $100 million worth of coins out of thin air.

iPhone 14 Pro chip bigger despite smaller transistors

Image: Circuit Board, Free Stock Picture, MorgueFile.com.
Closeup of old circuit board. Image: Circuit Board, Free Stock Picture, MorgueFile.com.

Small changes to caches and processor cores, this is how a preliminary analysis of the A16 silicon chip  from Angstronomics can be summarized. Although there is still no high-resolution image of the die, there is a video in which some details can already be seen. Since various components such as caches, processors and GPU form unique patterns on the die, they can be identified and at least roughly measured.

The operator of Angstronomics, who publishes under the pseudonym Skyjuice, comes to the conclusion that Apple has reduced the L3 or system level cache (SLC) in the A16 compared to the predecessor A15. Compared to the 4 MB L2 cache of the Efficiency CPU cluster made of Sawtooth cores, each of the two SLC blocks occupies about three times the area, so it should hold 12 MB – the SRAM memory cells need the same regardless of the cache hierarchy lots of space.

This means that the SLC of the A16, at 24 MB, is a quarter smaller than that of the A15, which has 32 MB. However, Apple has given the performance cores named Everest a third more L2 cache: the area here suggests that each of the two blocks holds 8 MB, while the A15 had a total of 12 MB.

One can only speculate about the reason for the reduction of the size of the SLC: Angronomics brings the higher data rate of the memory into play as a possible reason: LPDDR5-6400 is used for the first time in the A16. Optimizations are also conceivable, since the L2 cache of the P-cores was enlarged at the same time. Many factors play a role in the dimension of caches, including the micro architecture of the processors – it is very likely that there was not a single decisive argument for the redistribution.

Changes to processor cores

There are also small changes in the processor cores: they are arranged differently on the die, and Apple has also revised their structure. Both the Everest and Sawtooth (P/E) cores also appear to be slightly larger than their Avalanche and Blizzard predecessors. The neural and graphics processing units (NPU and GPU), on the other hand, seem to be quite unchanged. However, they are hardly recognizable in the Angtronomics image.

However, the NPU is only eight percent faster than the A15. This is part of the switch from the supplier TSMC from the N5 to the N4 process and the expected increase in speed of ten percent as a result. Major changes are therefore unlikely. The higher switching speed of the transistors in N4 should also play a role in the GPU, which also benefits from the larger memory bandwidth. Together, both could almost explain the measured 28 percent increase in speed .

Bigger chip despite (slightly) smaller transistors

With N4, TSMC refers to a further development of the N5 manufacturing process, with which Apple’s A15 is manufactured. According to TSMC, this increases the integration density by six percent, and the number of transistors also increases by six percent – ​​16 billion in the A16, 15 billion in the A15. Theoretically, the dies of A15 and A16 could be the same size.