Jeremy Howard
Jeremy Howard
  • Видео 188
  • Просмотров 8 711 177
The Fastlite DB library - Answer.AI dev chat #2
The 2nd in the series of informal chats amongst some of the Answer.AI R&D team. In this one, we introduce the basic ideas behind Fastlite (answerdotai.github.io/fastlite/ ), a new database library for Python.
Просмотров: 2 193

Видео

Claudette source walk-thru - Answer.AI dev chat #1
Просмотров 4 тыс.День назад
This is the 1st in a new series of informal chats amongst some of the Answer.AI R&D team. We will be recording these to help communicate progress within our team, but as a Public Benefit Corp with a mission to help society benefit from AI, that means it makes sense for us to share them with the rest of you folks too! We hope you enjoy this series. 😃
Going Further with CUDA for Python Programmers
Просмотров 11 тыс.4 месяца назад
This technical talk by Jeremy Howard explores advanced programming techniques for maximizing performance when using CUDA with Python. The focus is on optimizing memory usage with a specific emphasis on effectively leveraging fast shared memory in CUDA. It assumes you have already watched this "Getting Started" video: ruclips.net/video/nOxKexn3iBo/видео.html The video begins with foundational co...
Getting Started With CUDA for Python Programmers
Просмотров 51 тыс.5 месяцев назад
I used to find writing CUDA code rather terrifying. But then I discovered a couple of tricks that actually make it quite accessible. In this video I introduce CUDA in a way that will be accessible to Python folks, & I even show how to do it all for free in Colab! Notebooks This is lecture 3 of the "CUDA Mode" series (but you don't need to watch the others first). The notebook is available in th...
A Hackers' Guide to Language Models
Просмотров 509 тыс.9 месяцев назад
In this deeply informative video, Jeremy Howard, co-founder of fast.ai and creator of the ULMFiT approach on which all modern language models (LMs) are based, takes you on a comprehensive journey through the fascinating landscape of LMs. Starting with the foundational concepts, Jeremy introduces the architecture and mechanics that make these AI systems tick. He then delves into critical evaluat...
Jeremy Howard on ABC Weekend Breakfast
Просмотров 5 тыс.11 месяцев назад
Jeremy discusses the latest agreement between the US White House and US AI companies.
Jeremy Howard and Joshua Browder discuss AI & Jobs with Piers Morgan
Просмотров 8 тыс.Год назад
Full show link: ruclips.net/user/livegyhnHG5N2wE
Jeremy Howard demo for Mojo launch
Просмотров 131 тыс.Год назад
This is a section from the Modular launch video. The full video, docs, and details are here: www.modular.com/
Lesson 11 2022: Deep Learning Foundations to Stable Diffusion
Просмотров 21 тыс.Год назад
(All lesson resources are available at course.fast.ai.) In this lesson, we discuss various techniques and experiments shared by students on the forum, such as interpolating between prompts for visually appealing transitions and improving the update process in text-to-image generation, and a novel approach to decreasing the guidance scale during image generation. We then dive into a new paper ca...
Lesson 12: Deep Learning Foundations to Stable Diffusion
Просмотров 16 тыс.Год назад
(All lesson resources are available at course.fast.ai.) In this lesson, we start by discussing the CLIP Interrogator, a Hugging Face Spaces Gradio app that generates text prompts for creating CLIP embeddings. We then dive back into matrix multiplication, using Einstein summation notation and torch.einsum to simplify code and improve performance. We explore GPU acceleration with CUDA and Numba, ...
Lesson 13: Deep Learning Foundations to Stable Diffusion
Просмотров 15 тыс.Год назад
(All lesson resources are available at course.fast.ai.) In this lesson, we dive into backpropagation and the creation of a simple Multi-Layer Perceptron (MLP) neural network. We start by reviewing basic neural networks and their architecture, then move on to implementing a simple MLP from scratch. We focus on understanding the chain rule and backpropagation in the context of neural networks, an...
Lesson 14: Deep Learning Foundations to Stable Diffusion
Просмотров 13 тыс.Год назад
(All lesson resources are available at course.fast.ai.) In this lesson, we dive into the implementation of the chain rule in neural network training using backpropagation. We refactor our code to make it more efficient and flexible, and explore PyTorch's `nn.Module` and `nn.Sequential`. We also create custom PyTorch modules, build our own implementation of `nn.Module`, and learn about optimizer...
Lesson 15: Deep Learning Foundations to Stable Diffusion
Просмотров 11 тыс.Год назад
(All lesson resources are available at course.fast.ai.) We start with a dive into convolutional autoencoders and explore the concept of convolutions. Convolutions help neural networks understand the structure of a problem, making it easier to solve. We learn how to apply a convolution to an image using a kernel and discuss techniques like im2col, padding, and stride. We also create a CNN from s...
Lesson 16: Deep Learning Foundations to Stable Diffusion
Просмотров 9 тыс.Год назад
(All lesson resources are available at course.fast.ai.) In Lesson 16, we dive into building a flexible training framework called the learner. We start with a basic callbacks Learner, which is an intermediate step towards the flexible learner. We introduce callbacks, which are functions or classes called at specific points during the training process, and demonstrate the creation of a simple cal...
Lesson 17: Deep Learning Foundations to Stable Diffusion
Просмотров 8 тыс.Год назад
Lesson 17: Deep Learning Foundations to Stable Diffusion
Lesson 18: Deep Learning Foundations to Stable Diffusion
Просмотров 8 тыс.Год назад
Lesson 18: Deep Learning Foundations to Stable Diffusion
Lesson 19: Deep Learning Foundations to Stable Diffusion
Просмотров 9 тыс.Год назад
Lesson 19: Deep Learning Foundations to Stable Diffusion
Lesson 20: Deep Learning Foundations to Stable Diffusion
Просмотров 7 тыс.Год назад
Lesson 20: Deep Learning Foundations to Stable Diffusion
Lesson 21: Deep Learning Foundations to Stable Diffusion
Просмотров 7 тыс.Год назад
Lesson 21: Deep Learning Foundations to Stable Diffusion
Lesson 22: Deep Learning Foundations to Stable Diffusion
Просмотров 5 тыс.Год назад
Lesson 22: Deep Learning Foundations to Stable Diffusion
Lesson 23: Deep Learning Foundations to Stable Diffusion
Просмотров 6 тыс.Год назад
Lesson 23: Deep Learning Foundations to Stable Diffusion
Lesson 24: Deep Learning Foundations to Stable Diffusion
Просмотров 8 тыс.Год назад
Lesson 24: Deep Learning Foundations to Stable Diffusion
Lesson 25: Deep Learning Foundations to Stable Diffusion
Просмотров 6 тыс.Год назад
Lesson 25: Deep Learning Foundations to Stable Diffusion
Tanishq Mathew Abraham - Their Life and Work Eps 1
Просмотров 12 тыс.Год назад
Tanishq Mathew Abraham - Their Life and Work Eps 1
Lesson 10: Deep Learning Foundations to Stable Diffusion, 2022
Просмотров 59 тыс.Год назад
Lesson 10: Deep Learning Foundations to Stable Diffusion, 2022
Lesson 9B - the math of diffusion
Просмотров 31 тыс.Год назад
Lesson 9B - the math of diffusion
Lesson 9A 2022 - Stable Diffusion deep dive
Просмотров 31 тыс.Год назад
Lesson 9A 2022 - Stable Diffusion deep dive
Lesson 9: Deep Learning Foundations to Stable Diffusion
Просмотров 139 тыс.Год назад
Lesson 9: Deep Learning Foundations to Stable Diffusion
Fast.ai APL study session 17
Просмотров 2,9 тыс.Год назад
Fast.ai APL study session 17
Fast.ai APL study session 16
Просмотров 1,4 тыс.Год назад
Fast.ai APL study session 16

Комментарии

  • @codybontecou
    @codybontecou День назад

    Is there an updated url for the start_aws page?

    • @codybontecou
      @codybontecou День назад

      Scratch that. A quick google search of course fastai start aws shows the new url.

  • @kanewilliams1653
    @kanewilliams1653 2 дня назад

    A lot to think about..!

  • @tecnom7133
    @tecnom7133 4 дня назад

    Thanks alot :)

  • @SarisKiattithapanayong-hx3dw
    @SarisKiattithapanayong-hx3dw 4 дня назад

    how can you make MAC book using CUDa?

  • @rubensmau
    @rubensmau 5 дней назад

    Many thanks, very useful

  • @MichaelChenAdventures
    @MichaelChenAdventures 5 дней назад

    Thank you Jeremy!

  • @alexkelly757
    @alexkelly757 6 дней назад

    Where are people from answer AI posting videos? is it under there own account, just wondering where to watch out for new videos

    • @fredguth1315
      @fredguth1315 6 дней назад

      They are using Jeremy's account which already has more than 100k subscribers. Same with Fastai.

  • @Little-bird-told-me
    @Little-bird-told-me 6 дней назад

    this one presentation is worth more than all the AI discourse on internet.

  • @user-kn4wt
    @user-kn4wt 6 дней назад

    duckdb is king. not sure about this

  • @marekglowacki2607
    @marekglowacki2607 6 дней назад

    There is a small typo in link to fastlite (right parenthesis included)

  • @wolpumba4099
    @wolpumba4099 6 дней назад

    *Summary* *Key takeaways:* * *Fastlite* is a new Python database library built on top of Simon Willison's `sqlite-utils`. (0:08) * It focuses on *ergonomics and ease of use* within interactive environments like Jupyter notebooks and IPython. (12:59) * *Key features:* * *Dynamic tab completion:* Access tables, views, and columns with tab completion. (11:12) * *Simplified syntax:* Use intuitive dot notation (e.g., `db.t.artists`) for accessing database elements. (13:59) * *Built-in visualization:* Generate database schema diagrams directly within the notebook using Graphviz. (22:45) * *Leverages Python's dynamic nature:* Utilizes magic methods like `__repr__` and `__getattr__` for a more Pythonic experience. (30:30) * *Motivation:* * Simplify web application development by making database interactions more intuitive. (0:48) * Address the complexity and verbosity of existing ORMs like SQLAlchemy and SQLModel. (1:16) * Provide a more dynamic and interactive alternative to statically typed approaches. (9:18) * *Implementation:* * Leverages `sqlite-utils` for core database functionality. (5:02) * Uses Python's magic methods to enable tab completion and custom object representations. (33:35) * Employs Graphviz for generating database schema diagrams. (47:32) * *Benefits:* * Easier to learn and use, especially for those familiar with SQL. * More interactive and exploratory workflow. * Less boilerplate code compared to traditional ORMs. *Overall:* Fastlite aims to make working with databases in Python more intuitive and enjoyable, particularly within interactive environments. i used gemini 1.5 pro to summarize the transcript

  • @Little-bird-told-me
    @Little-bird-told-me 6 дней назад

    Can someone tell me if I need to buy an Nvidia GPU to run ML task ? I wanted to get an AMD but since they don't support CUDA(i may be wrong) I am a little apprehensive to go with it.

  • @AyahuascaDataScientist
    @AyahuascaDataScientist 6 дней назад

    Glad Jeremy is posting more. If you see this Jeremy…I have one message for you: POST MORE! Share your knowledge!

  • @arnokhachatourian8928
    @arnokhachatourian8928 6 дней назад

    Auto-diagraming seems pretty clean! also, heads up that there's a sneaky parens `)` on the end of your link that's breaking it.

  • @fredguth1315
    @fredguth1315 6 дней назад

    Another interesting idea is using duckdb python api.

  • @davidz6828
    @davidz6828 7 дней назад

    Thanks for sharing, Jeremy and team. Always learn something new from you.

  • @niazhimselfangels
    @niazhimselfangels 9 дней назад

    Only fifteen minutes in and I'm already missing fastai classes so much! There's a ton I learn from you Jeremy, whenever you do one of these streams. And the Hackers Guide to LLMs video is a true gem! 🎉

  • @law_wu
    @law_wu 9 дней назад

    Thanks for this Jeremy and team! So informative to learn how these libraries are built. Looking forward to future dev chats.

  • @souvikbhattacharyya2480
    @souvikbhattacharyya2480 9 дней назад

    Thanks for making these public!

  • @pkn8707
    @pkn8707 9 дней назад

    Until and unless, educators like Jeremy are present, no closed source company can have a lock on knowledge. Thanks for doing what you do, so consistently. One question though, even though there's so much chaos in education field, what motivates you to do it consistently? Doing great is okay, doing great consistently is really hard in this distraction prone world. Anyways as always Thank you and your team for your contribution

  • @EkShunya
    @EkShunya 9 дней назад

    great video as usual

  • @wolpumba4099
    @wolpumba4099 9 дней назад

    *Summary* *Why Claudette exists:* * *[**0:00**]* Jeremy feels Claude is underrated and wants to promote its use. * *[**4:03**]* He aims to provide a simpler, more transparent alternative to large, complex LLM frameworks. * *[**6:18**]* Claudette bridges the gap between writing everything from scratch and using bulky frameworks, appealing to both beginners and experienced Python programmers. *What Claudette does:* * *Simplifies Anthropic SDK:* * *[**8:27**]* Offers convenient functions to access Claude's models (Opus, Sonnet, Haiku). * *[**15:57**]* Handles message formatting and content extraction, making interactions cleaner. * *[**19:15**]* Tracks token usage across sessions for cost monitoring. * *[**24:48**]* Provides easy ways to use `prefill` (forcing the start of a response) and streaming. * *Implements Chat History:* * *[**1:01:55**]* The `Chat` class maintains conversation history, mimicking stateful behavior. * *[**1:04:34**]* Integrates seamlessly with prefill, streaming, and tool use within a chat session. * *Facilitates Tool Use (REACT pattern):* * *[**36:57**]* Uses Python functions as tools, automatically generating the required JSON schema with `docstrings` and type hints. * *[**56:21**]* Handles tool execution based on Claude's requests, including passing inputs and retrieving results. * *[**1:28:45**]* Provides a `tool_loop` function to automate multi-step tool interactions. * *Supports Images:* * *[**1:09:01**]* Includes functions to easily incorporate images into messages and prompts. * *Example Use Cases:* * *[**1:19:12**]* Demonstrates building a simple customer service agent with tool use. * *[**1:31:20**]* Showcases building a code interpreter similar to ChatGPT's, combining Python execution with other tools. *Key Features:* * *[**7:36**]* *Literate Programming:* The source code is designed to be read like a tutorial, explaining itself as you go. * *[**5:09**]* *Minimal Abstractions:* Leverages existing Python knowledge and avoids introducing unnecessary complexity. * *[**1:24:51**]* *Transparency:* You can easily inspect requests, responses, history, and even debug the HTTP communication. *Future Plans:* * *[**1:16:19**]* Create similar libraries ("friends") for other LLMs like GPT and Gemini. * *[**1:16:19**]* Maintain focus on simplicity and ease of use. *Overall:* Claudette is a user-friendly library that simplifies working with Claude while providing powerful features for building LLM applications. Its literate programming style and minimal abstractions make it easy to learn, use, and extend. i used gemini 1.5 pro to summarize the transcript.

  • @user-go4sb6re4y
    @user-go4sb6re4y 9 дней назад

    Thanks for podcast!

  • @bloodofwebseries6584
    @bloodofwebseries6584 11 дней назад

    What basic concept i have to know to understand this course

  • @giorda77
    @giorda77 12 дней назад

    Lesson 11 and 12 should be called "Broadcasting:the missing semester." Great content @Jeremy Howard. Thanks again

  • @adityagupta-hm2vs
    @adityagupta-hm2vs 13 дней назад

    Also, are we using latent space as gradients here, as we are subtracting gradients from the latent, which we typically do from weights in conventional NN ?

  • @adityagupta-hm2vs
    @adityagupta-hm2vs 13 дней назад

    How do we decide the scaling factor in VAE part i.e. 0.18215, any hint on how to decide it ? I did try changing and could see the different output, but what's a good way to choose ?

  • @mchristos
    @mchristos 13 дней назад

    Q: why not just make bs=len(valid_ds), i.e. make the batch size for the validation set the same as its length? I can't see a function in having batches of the validation set, since we're just computing some metric on it?

  • @paulsawyer1288
    @paulsawyer1288 16 дней назад

    Did you ever use your understanding to teach your peers in classes?

  • @davisisibor3509
    @davisisibor3509 20 дней назад

    Watching this last few sections of this video as a frontend developer . Was one of my best moments 😂😂😂😂. I skipped the entire thing 😂😂😂😂😂

  • @nbn_keramoti
    @nbn_keramoti 23 дня назад

    A hackers' guide to using language models, including open-source and OpenAI models, with a focus on a code-first approach. Covers language model pre-training, fine-tuning, and reinforcement learning from human feedback. Demonstrates creating a custom code interpreter and fine-tuning a model for SQL generation. Key moments: 00:01 Language models are essential in predicting the next word or filling in missing words in a sentence. They use tokens, which can be whole words or subword units, and are trained through pre-training on large datasets like Wikipedia. -Language models predict the next word or fill in missing words. They use tokens that can be whole words or subword units, enhancing their predictive capabilities. -Training language models involves pre-training on extensive datasets like Wikipedia. This process helps the model learn language patterns and improve its predictive accuracy. 08:04 Neural networks, specifically deep neural networks, are trained to predict the next word in a sentence by learning about the world and building abstractions. This process involves compression and fine-tuning through language model fine-tuning and classifier fine-tuning. -The importance of neural networks learning about objects, time, movies, directors, and people to predict words effectively. This knowledge is crucial for language models to perform well in various tasks. -The concept of compression in neural networks and the relationship between compression and intelligence. Fine-tuning through language model fine-tuning and classifier fine-tuning enhances the model's capabilities. -Different approaches like instruction tuning and reinforcement learning from human feedback are used in classifier fine-tuning to improve the model's performance in answering questions and solving problems. 16:07 To effectively use language models, starting with being a proficient user is crucial. GPT-4 is currently recommended for language modeling, offering capabilities beyond common misconceptions about its limitations. -GPT-4's ability to address reasoning challenges and common misconceptions about its limitations. It can effectively solve problems when primed with custom instructions. -The training process of GPT-4 and the importance of understanding its initial purpose to provide accurate answers. Custom instructions can guide GPT-4 to offer high-quality information. -The impact of custom instructions on GPT-4's problem-solving capabilities and the ability to generate accurate responses by priming it with specific guidelines. 24:12 Language models like GPT-4 can provide concise answers but may struggle with self-awareness and complex logic puzzles, leading to hallucinations and errors. Encouraging multiple attempts and using advanced data analysis can improve accuracy. -Challenges with self-awareness and complex logic puzzles can lead to errors and hallucinations in language models like GPT-4, affecting the accuracy of responses. -Encouraging multiple attempts and utilizing advanced data analysis can enhance the accuracy of language models like GPT-4 in providing solutions to complex problems. -Utilizing advanced data analysis allows for requesting code generation and testing, improving efficiency and accuracy in tasks like document formatting. 32:20 Language models like GPT-4 excel at tasks that require familiarity with patterns and data processing, such as extracting text from images, creating tables, and providing quick responses based on predefined instructions. -The efficiency of language models in tasks like extracting text from images and creating tables due to their ability to recognize patterns and process data quickly. -Comparison between GPT-4 and GPT 3.5 in terms of cost-effectiveness for using the Open AI API, showcasing the affordability and versatility of GPT models for various tasks. -The practical applications of using the Open AI API programmatically for data analysis, repetitive tasks, and creative programming, offering a different approach to problem-solving. 40:26 Understanding the cost and usage of OpenAI's GPT models is crucial. Monitoring usage, managing rate limits, and creating custom functions can enhance the experience and efficiency of using the API. -Monitoring usage and cost efficiency of OpenAI's GPT models is essential to avoid overspending. Testing with lower-cost options before opting for expensive ones can help in decision-making. -Managing rate limits is important when using OpenAI's API. Keeping track of usage, especially during initial stages, and implementing functions to handle rate limit errors can prevent disruptions in service. -Creating custom functions and tools can enhance the functionality of OpenAI's API. Leveraging function calling and passing keyword arguments can enable the development of personalized code interpreters and utilities. 48:30 Creating a code interpreter using GPT-4 allows for executing code and returning results. By building functions, one can enhance the model's capabilities beyond standard usage. -Exploring the concept of doc strings as the key for programming GPT-4, highlighting the importance of accurate function descriptions for proper execution. -Utilizing custom functions to prompt GPT-4 for specific computations, showcasing the model's ability to determine when to use provided functions. -Enhancing GPT-4's functionality by creating a Python function for executing code and returning results, ensuring security by verifying code before execution. 56:34 Using Fast AI allows accessing others' computers for cheaper and better availability. GTX 3090 is recommended for language models due to memory speed over processor speed. -Options for renting GPUs include GTX 3090 for $700 or A6000 for $5000, with memory size considerations. Using a Mac with M2 Ultra can be an alternative. -Utilizing the Transformers library from Hugging Face for pre-trained models. Challenges with model evaluation metrics and potential data leakage in training sets. -Selecting models based on Metas Llama 2 for language models. Importance of fine-tuning pre-trained models for optimal performance and memory considerations. 1:04:38 Jeremy Howard, an Australian AI researcher and entrepreneur, discusses optimizing language models for speed and efficiency by using different precision data formats, such as B float 16 and gptq, resulting in significant time reductions. -Exploring the use of B float 16 and gptq for optimizing language models, leading to faster processing speeds and reduced memory usage. -Utilizing instruction-tuned models like stable Beluga and understanding the importance of prompt formats during the instruction tuning process. -Implementing retrieval augmented generation to enhance language model responses by searching for relevant documents like Wikipedia and incorporating the retrieved information. 1:12:42 The video discusses using open-source models with context lengths of 2000-4000 tokens to answer questions by providing context from web pages. It demonstrates using a sentence Transformer model to determine the most relevant document for answering a question. -Utilizing sentence Transformer models to identify the most suitable document for answering questions based on similarity calculations. -Exploring the process of encoding documents and questions to generate embeddings for comparison and selecting the most relevant document. -Discussing the use of vector databases for efficient document encoding and retrieval in large-scale information processing tasks. 1:20:46 Fine-tuning models allows for customizing behavior based on available documents, demonstrated by creating a tool to generate SQL queries from English questions, showcasing the power of personalized model training in just a few hours. -Utilizing the Hugging Face data sets library for fine-tuning models, enabling quick customization based on specific datasets for specialized tasks. -Exploring the use of Axolotl, an open-source software, to fine-tune models efficiently, showcasing the ease of implementation and ready-to-use functionalities for model training. -Discussing alternative options for model training on Mac systems, highlighting the mlc and llama.cpp projects that offer flexibility in running language models on various platforms. 1:28:50 Exploring language models like Llama can be exciting yet challenging for Python programmers due to rapid development, early stages, and installation complexities. -Benefits of using Nvidia graphics card and being a capable Python programmer for utilizing Pi torch and hugging face ecosystem in language model development. -The evolving nature of language models like Llama, the abundance of possibilities they offer, and the importance of community support through Discord channels.

  • @marko.p.radojcic
    @marko.p.radojcic 24 дня назад

    I am getting RUclips premium just! to download this series. Thank you!

  • @adnanwahab4191
    @adnanwahab4191 Месяц назад

    Amazing content thank you so much !

  • @shubh9207
    @shubh9207 Месяц назад

    Please roll out the next series, although I'm in the first part, I just can't wait to reach here and learn from such amazing tutors.

  • @DeltaJes-co8yu
    @DeltaJes-co8yu Месяц назад

    Kerala has something to be proud about!

  • @shubh9207
    @shubh9207 Месяц назад

    Hello, is it fine to run Jupyter notebook on VS code instead of running it on a local browser?

  • @joxa6119
    @joxa6119 Месяц назад

    What does Jeremy throw to the audience? is it mic

  • @shadril2383
    @shadril2383 Месяц назад

    best playlist for absolute beginners!

  • @KetanSingh
    @KetanSingh Месяц назад

    I try watching this once every year. Incredibly good course

  • @DinoFancellu
    @DinoFancellu Месяц назад

    Don't like all this jumping around. Would be much easier to simply go through it, in a linear fashion, explaining as you go. Disappointing

  • @frankchieng
    @frankchieng Месяц назад

    i thought in the class of WandBCB(MetricsCB) def _log(self, d): if self.train: should be modified with if d['train']=='train'

  • @bbalban
    @bbalban Месяц назад

    great course! so weird that the videos have less than 100k views.

  • @RayhaanKhan-mu4qu
    @RayhaanKhan-mu4qu Месяц назад

    29:45 I can agree that anime people watch wayy too much anime!

  • @swimmingpolar
    @swimmingpolar Месяц назад

    First comment on RUclips here. Among all those videos on RUclips, using custom instruction like what you did is literally eye opening. I thought current AI models’ limitations are limited by nature that it can’t be improved. Of course it is that you are professional in AI but things are so organized well and straightforward that I can understand and see the result right away. 😂 Gonna have to steal your instruction as well.

  • @bilalch83
    @bilalch83 Месяц назад

    Sensei Jeremy Howard

  • @bayesianmonk
    @bayesianmonk Месяц назад

    Sometimes explaining the math helps more than escaping it, no heavy math is used anyway. I found the explanation of DDIM not very clear. Thanks for the course and videos.

    • @thehigheststateofsalad
      @thehigheststateofsalad Месяц назад

      We need another session to explain this process.

    • @maxkirby8500
      @maxkirby8500 25 дней назад

      Yeah. I've been spending quite a bit of time trying to bridge the gap by reading through the papers and stuff, but maybe that's intented...

  • @antonioalvarado7594
    @antonioalvarado7594 Месяц назад

    Freedom for deep learning: Unlocked. Thank you sir.

  • @user-kl1dc8nh3l
    @user-kl1dc8nh3l Месяц назад

    really great content.

  • @ItzGanked
    @ItzGanked Месяц назад

    jupyter also has documentation for python and other libraries inside the notebook if you go to help python reference. Keeps you inside of the notebook instead of searching the web and will provide docs for the version of python you are using.

  • @alecmorgan3541
    @alecmorgan3541 Месяц назад

    You wouldn't download 600 images of bears!