Subscribe for only $15 to access all of our content
Assigns allow you to set and access shared data in different contexts within Phoenix. In this episode we’ll explore how to use assigns in the connection, template, and socket.
In this episode we’ll learn how we can use feature flags to toggle different features of an application to different users. To do this we’ll be using the FunWithFlags package in an existing Elixir application.
In this episode we’ll look at how we can use the Phoenix router to define routes for our application. We’ll learn how to define single routes and full RESTful resources in Phoenix.
In this episode we’ll update an existing Elixir application to make it use a Stream throughout. Then we’ll see how we can use Task.async_stream to make concurrent API calls.
In this episode we’ll learn the basics of working with CSV files in Elixir. We’ll parse and existing CSV to get specific data from it. We’ll also learn how to build a new CSV file.
In this episode we’ll learn how to seed our Ecto database with some dummy info for use in a Phoenix application.
In this episode we’ll learn how we can use virtual attributes in Ecto to create fields in a schema that are not persisted to the database.
In this episode we’ll create a new Mix project to interact with an FTP server. We’ll learn how to connect to the server, change directories, download and upload files.
In this episode we’ll look at how we can implement passwordless authentication in a Phoenix application with the Veil package.
Ready to dig into Supervisors? In this episode we’ll get introduced to them by updating an existing supervisor and its corresponding worker to use the Elixir 1.5 syntax.
In this episode we’ll add user authentication to a Phoenix application. We’ll create forms that allow users to sign up and sign in. Then we’ll see how we can require users to authenticate before accessing certain controller actions.
If you’ve needed to upload and transform images in your Phoenix application this episode’s for you. We’ll use Arc to create multiple versions of our files and upload them to Amazon S3.