cookie

Ми використовуємо файли cookie для покращення вашого досвіду перегляду. Натиснувши «Прийняти все», ви погоджуєтеся на використання файлів cookie.

avatar

TechPulse: AI & Innovation

Providing you with the latest updates and insights in technology, artificial intelligence, physics, and astronomy. Join us on a journey to explore the frontiers of knowledge and imagination! Support : @TechPuls_bot

Більше
Рекламні дописи
2 786
Підписники
-324 години
-257 днів
-19430 днів

Триває завантаження даних...

Приріст підписників

Триває завантаження даних...

"Time Travel is now simply an engineering problem" - Michio Kaku
Показати все...
Фото недоступнеДивитись в Telegram
Taking care of yourself is not selfish, it’s the key to happiness. Making your own happiness a top priority allows you to care for those around you more deeply. Remember: you can’t pour from an empty cup. So do something today that recharges you. ❤️ Kristen Butler
Показати все...
3
Sometimes the smallest step in the right direction ends up being the biggest step of your life. There is no way to know what life has in store for us, but these small acts of faith can certainly lead us to where we’re headed. Remember, life is a sum of small decisions and actions. So trust your mind, heart and gut and take that step. Life is full of opportunities; it is up to you to grab them. Jim Kwik
Показати все...
Repost from Du Rove's Channel
🇺🇿 This week I visited Uzbekistan – and thoroughly enjoyed it 👍 ❤️ Uzbekistan loves Telegram: over 70% of the country’s 37 million people is on Telegram and their entire economy is run on our platform (every business in the country has a Telegram bot or channel). We are proud of this popularity and we love Uzbekistan back ❤️ ✨During my trip to Uzbekistan I was amazed by the modern infrastructure of quickly developing Tashkent, the mountainous landscapes of Eastern Uzbekistan, and the rich history of Buhara and Samarkand 💖 🏄‍♂️ I’ve met lots of young and talented people with good hearts. I am grateful for the warm welcome and hope to stay more in the sunny Tashkent later this year 😎
Показати все...
👍 2 1
An artificial intelligence candidate has officially entered the race for a seat in the UK Parliament. This historic development marks the first time an Al entity has put forth its candidacy for elected office in Britain. While the notion of an Al running for Parliament may seem like something out of science fiction, the candidate's backers argue that an artificial intelligence could bring impartiality, data-driven decision making, and a long-term perspective unencumbered by human limitations to the role. Regardless of the outcome, this Al's campaign will no doubt spur serious debate about the role artificial intelligence may play in the future of governance,
Показати все...
👍 1
00:31
Відео недоступнеДивитись в Telegram
9.34 MB
2
AR Laptop
Показати все...
Sightful has introduced the Spacetop G1, an augmented reality laptop with a streamlined design and integrated AR glasses, runs on SpaceOS and a Qualcomm Snapdragon QCS8550. The device, projected to ship in October for $1700, aims to enhance productivity with a 100" virtual screen.
Показати все...
6.26 MB
1👏 1
Показати все...
Let's reproduce GPT-2 (124M)

We reproduce the GPT-2 (124M) from scratch. This video covers the whole process: First we build the GPT-2 network, then we optimize its training to be really fast, then we set up the training run following the GPT-2 and GPT-3 paper and their hyperparameters, then we hit run, and come back the next morning to see our results, and enjoy some amusing model generations. Keep in mind that in some places this video builds on the knowledge from earlier videos in the Zero to Hero Playlist (see my channel). You could also see this video as building my nanoGPT repo, which by the end is about 90% similar. Links: - build-nanogpt GitHub repo, with all the changes in this video as individual commits:

https://github.com/karpathy/build-nanogpt

- nanoGPT repo:

https://github.com/karpathy/nanoGPT

- llm.c repo:

https://github.com/karpathy/llm.c

- my website:

https://karpathy.ai

- my twitter:

https://twitter.com/karpathy

- our Discord channel:

https://discord.gg/3zy8kqD9Cp

Supplementary links: - Attention is All You Need paper:

https://arxiv.org/abs/1706.03762

- OpenAI GPT-3 paper:

https://arxiv.org/abs/2005.14165

- OpenAI GPT-2 paper:

https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf-

The GPU I'm training the model on is from Lambda GPU Cloud, I think the best and easiest way to spin up an on-demand GPU instance in the cloud that you can ssh to:

https://lambdalabs.com

Chapters: 00:00:00 intro: Let’s reproduce GPT-2 (124M) 00:03:39 exploring the GPT-2 (124M) OpenAI checkpoint 00:13:47 SECTION 1: implementing the GPT-2 nn.Module 00:28:08 loading the huggingface/GPT-2 parameters 00:31:00 implementing the forward pass to get logits 00:33:31 sampling init, prefix tokens, tokenization 00:37:02 sampling loop 00:41:47 sample, auto-detect the device 00:45:50 let’s train: data batches (B,T) → logits (B,T,C) 00:52:53 cross entropy loss 00:56:42 optimization loop: overfit a single batch 01:02:00 data loader lite 01:06:14 parameter sharing wte and lm_head 01:13:47 model initialization: std 0.02, residual init 01:22:18 SECTION 2: Let’s make it fast. GPUs, mixed precision, 1000ms 01:28:14 Tensor Cores, timing the code, TF32 precision, 333ms 01:39:38 float16, gradient scalers, bfloat16, 300ms 01:48:15 torch.compile, Python overhead, kernel fusion, 130ms 02:00:18 flash attention, 96ms 02:06:54 nice/ugly numbers. vocab size 50257 → 50304, 93ms 02:14:55 SECTION 3: hyperpamaters, AdamW, gradient clipping 02:21:06 learning rate scheduler: warmup + cosine decay 02:26:21 batch size schedule, weight decay, FusedAdamW, 90ms 02:34:09 gradient accumulation 02:46:52 distributed data parallel (DDP) 03:10:21 datasets used in GPT-2, GPT-3, FineWeb (EDU) 03:23:10 validation data split, validation loss, sampling revive 03:28:23 evaluation: HellaSwag, starting the run 03:43:05 SECTION 4: results in the morning! GPT-2, GPT-3 repro 03:56:21 shoutout to llm.c, equivalent but faster code in raw C/CUDA 03:59:39 summary, phew, build-nanogpt github repo Corrections: I will post all errata and followups to the build-nanogpt GitHub repo (link above) SuperThanks: I experimentally enabled them on my channel yesterday. Totally optional and only use if rich. All revenue goes to to supporting my work in AI + Education.

👏 2
Оберіть інший тариф

На вашому тарифі доступна аналітика тільки для 5 каналів. Щоб отримати більше — оберіть інший тариф.