일 | 월 | 화 | 수 | 목 | 금 | 토 |
---|---|---|---|---|---|---|
1 | 2 | 3 | 4 | |||
5 | 6 | 7 | 8 | 9 | 10 | 11 |
12 | 13 | 14 | 15 | 16 | 17 | 18 |
19 | 20 | 21 | 22 | 23 | 24 | 25 |
26 | 27 | 28 | 29 | 30 | 31 |
- Absolute
- AGI
- ai
- AI agents
- AI engineer
- AI researcher
- ajax
- algorithm
- Algorithms
- aliases
- Array 객체
- ASI
- bayes' theorem
- Bit
- Blur
- BOM
- bootstrap
- canva
- challenges
- ChatGPT
- Today
- In Total
목록전체 글 (553)
A Joyful AI Research Journey🌳😊
The "Attention Is All You Need" paper by Vaswani et al. (2017) introduced the Transformer architecture, a groundbreaking method in Natural Language Processing. Access the paper and its PDF version through the following links: Read the Paper on arXiv https://arxiv.org/abs/1706.03762 Attention Is All You Need The dominant sequence transduction models are based on complex recurrent or convolutional..
ChatGPT, response to “Is it better to first understand seq2seq models in-depth and then use high-level libraries like Hugging Face or TensorFlow? Is this approach similar to studying theory first and then using a library?,” August 27, 2023, OpenAI. Yes, your understanding is on point. Let's delve into why this sequential approach of starting with seq2seq and then moving on to modern libraries li..
ChatGPT, OpenAI. The distinction between an "AI engineer" and an "AI researcher" often lies in the nature of the work and the environment, although the boundaries can sometimes be blurred: AI Engineer: Typically works in the industry, applying established AI and machine learning techniques to real-world problems. Might develop and implement machine learning models, optimize existing algorithms f..
ChatGPT, response to “Can I use a seq2seq model for NMT using Hugging Face, Keras, or TensorFlow?” August 27, 2023, OpenAI. Yes, the seq2seq (sequence-to-sequence) model is a foundational architecture for NMT (Neural Machine Translation), and you can implement and train it using any of the mentioned frameworks: Hugging Face's Transformers, Keras, or TensorFlow. Here's a brief overview of how you..
As I continue to develop and refine my work in the realm of NLP and AI, some sub-categories might appear empty. This is because I've chosen to temporarily set certain posts to private. I'm doing this to ensure the quality and accuracy of the content I present to you. Rest assured, updates and detailed insights will be made public in due course.
I'd like to inform you that the sub-categories related to my postgraduate study are currently empty. The content within these categories, except for the lectures I am going to take, is of a personal and confidential nature. I've chosen to keep these documents private to protect my personal information. I appreciate your understanding and respect for my privacy. Thank you for your continued suppo..
I recently encountered an in-depth tutorial by Andrej Karpathy regarding the construction of the GPT model from its foundational principles. For individuals interested in a comprehensive understanding of the GPT architecture and its mechanisms, I highly recommend viewing the presentation titled, "Let's build GPT: from scratch, in code, spelled out." Happy learning! :) https://www.youtube.com/wat..
If you're interested in the latest advancements in coding language models, you should definitely look into Code Llama. It's a cutting-edge large language model designed specifically for coding: Introducing Code Llama: Discover the unique features and capabilities of Code Llama, an enhanced version of Llama 2 specialized in code. Read more on Meta AI's blog post. Code Llama Research Paper: Dive d..
ChatGPT, response to “Is NMT different from MT?,” August 24, 2023, OpenAI. NMT stands for "Neural Machine Translation," which is a subfield of MT (Machine Translation). NMT specifically refers to machine translation approaches based on deep learning and neural networks, while MT can refer to any automated method of translating text or speech from one language to another, including older methods ..