diff --git a/.gitea/workflows/tests.yaml b/.gitea/workflows/tests.yaml
index a906e41..860e0b7 100644
--- a/.gitea/workflows/tests.yaml
+++ b/.gitea/workflows/tests.yaml
@@ -21,7 +21,7 @@ jobs:
   # our workaround is to reference service containers two different ways.
   #   - when running nektos/act on localhost, we reference the service as localhost:<port_number>.
   #   - when running gitea act_runner on gitea, we reference the service's dns name.
-
+  # these references are defined in environment variables
   test_phoenix:
     env:
       SECRET_KEY_BASE: ${{ secrets.SECRET_KEY_BASE }}
diff --git a/MANTRAS.md b/MANTRAS.md
index ebee7de..7bb669a 100644
--- a/MANTRAS.md
+++ b/MANTRAS.md
@@ -50,4 +50,9 @@ In other words, pick something for a name and roll with the punches.
 > "But it's also about looking at things anew and what they could be instead of what they are"
 > -- Rodney Mullen
 
-### Success requires continuing even when it's painful
\ No newline at end of file
+### Success requires continuing even when it's painful
+
+### Find what you love and let it kill you
+
+> Find what you love and let it kill you. Let it drain you of your all. Let it cling onto your back and weigh you down into eventual nothingness. Let it kill you and let it devour your remains. For all things will kill you, both slowly and fastly, but it's much better to be killed by a lover.
+> -- Charles Bukowski
\ No newline at end of file
diff --git a/ansible/requirements.yml b/ansible/requirements.yml
index 88abd9b..6fa5fd5 100644
--- a/ansible/requirements.yml
+++ b/ansible/requirements.yml
@@ -3,5 +3,6 @@ collections:
   - name: cloud.terraform
   - name: community.docker
   - name: community.general
+  - name: gluster.gluster
 roles:
   - name: nvjacobo.caddy
diff --git a/ansible/roles/capture/tasks/main.yml b/ansible/roles/capture/tasks/main.yml
index d4269f4..73a11ad 100644
--- a/ansible/roles/capture/tasks/main.yml
+++ b/ansible/roles/capture/tasks/main.yml
@@ -54,7 +54,7 @@
 
 - name: Download and extract IPFS Kubo
   ansible.builtin.unarchive:
-    src: https://dist.ipfs.tech/kubo/v0.33.2/kubo_v0.33.2_linux-amd64.tar.gz
+    src: "https://dist.ipfs.tech/kubo/{{ kubo_version }}/kubo_{{ kubo_version }}_linux-amd64.tar.gz"
     dest: ~/
     remote_src: true
 
@@ -63,12 +63,15 @@
   args:
     creates: /usr/local/bin/ipfs
 
+
+
+
 - name: Allow UFW ports
   community.general.ufw:
     rule: allow
     port: "{{ item }}"
     proto: tcp
   loop:
-    - 8081 # npx http-server -p 8081 
+    - 8081 # npx http-server -p 8081
     - 8080 # ipfs api
     - 4001 # ipfs swarm
diff --git a/ansible/roles/ipfs/defaults/main.yml b/ansible/roles/ipfs/defaults/main.yml
new file mode 100644
index 0000000..06f8223
--- /dev/null
+++ b/ansible/roles/ipfs/defaults/main.yml
@@ -0,0 +1,3 @@
+---
+ipfs_kubo_version: v0.34.1
+ipfs_cluster_follow_version: v1.1.2
diff --git a/ansible/roles/ipfs/handlers/main.yml b/ansible/roles/ipfs/handlers/main.yml
new file mode 100644
index 0000000..7835f14
--- /dev/null
+++ b/ansible/roles/ipfs/handlers/main.yml
@@ -0,0 +1,16 @@
+---
+
+- name: Restart ipfs
+  ansible.builtin.systemd_service:
+    name: ipfs
+    state: restarted
+    enabled: true
+    daemon_reload: true
+
+- name: Restart ipfs-cluster-follow
+  ansible.builtin.systemd_service:
+    name: ipfs-cluster-follow
+    state: restarted
+    enabled: true
+    daemon_reload: true
+
diff --git a/ansible/roles/ipfs/tasks/main.yml b/ansible/roles/ipfs/tasks/main.yml
new file mode 100644
index 0000000..b0e7ef4
--- /dev/null
+++ b/ansible/roles/ipfs/tasks/main.yml
@@ -0,0 +1,175 @@
+---
+- name: Configure firewall (UDP & TCP)
+  community.general.ufw:
+    rule: allow
+    port: "{{ item }}"
+    proto: any
+  loop:
+    - 4001
+    - 24007
+    - 24008
+
+- name: Configure firewall (TCP)
+  community.general.ufw:
+    rule: allow
+    port: "{{ item }}"
+    proto: tcp
+  loop:
+    - 29152:65535
+
+
+- name: Install glusterfs
+  ansible.builtin.apt:
+    name:
+      - glusterfs-server
+    state: present
+
+- name: Start & enable glusterd service
+  ansible.builtin.systemd_service:
+    name: glusterd.service
+    state: started
+    enabled: true
+
+# - name: Create gluster volume
+#   gluster.gluster.gluster_volume:
+#     state: present
+#     name: ipfs-datastore
+#     bricks: /bricks/brick1/g1
+#     rebalance: true
+#     cluster: "{{ groups['ipfs'] }}"
+#   run_once: true
+
+
+# - name: Start gluster volume
+#   gluster.gluster.gluster_volume:
+#     state: started
+#     name: ipfs-datastore
+
+# - name: Limit volume usage
+#   gluster.gluster.gluster_volume:
+#     state: present
+#     name: ipfs-datastore
+#     directory: /
+#     quota: 6.0TB
+
+## Example: mount -t glusterfs fp-bright-0:/gv0 /mountme
+# - name: Mount gluster volume
+#   ansible.posix.mount:
+#     src: "{{ ansible_hostname }}:/g1"
+#     path: /mnt/g1
+#     fstype: glusterfs
+#     state: mounted
+
+- name: Create ipfs group
+  ansible.builtin.group:
+    name: ipfs
+    state: present
+
+- name: Create ipfs user
+  ansible.builtin.user:
+    name: ipfs
+    group: ipfs
+    create_home: true
+    home: /home/ipfs
+    system: true
+
+- name: Download and extract IPFS Kubo
+  ansible.builtin.unarchive:
+    src: "https://dist.ipfs.tech/kubo/{{ ipfs_kubo_version }}/kubo_{{ ipfs_kubo_version }}_linux-amd64.tar.gz"
+    dest: /tmp
+    remote_src: true
+  notify:
+    - Restart ipfs
+
+- name: Install IPFS Kubo
+  ansible.builtin.copy:
+    src: /tmp/kubo/ipfs
+    dest: /usr/local/bin/ipfs
+    mode: "0755"
+    remote_src: true
+  notify:
+    - Restart ipfs
+
+- name: Download and extract ipfs-cluster-follow
+  ansible.builtin.unarchive:
+    src: "https://dist.ipfs.tech/ipfs-cluster-follow/{{ ipfs_cluster_follow_version }}/ipfs-cluster-follow_{{ ipfs_cluster_follow_version }}_linux-amd64.tar.gz"
+    dest: /tmp
+    remote_src: true
+  notify:
+    - Restart ipfs-cluster-follow
+
+- name: Install ipfs-cluster-follow
+  ansible.builtin.copy:
+    src: /tmp/ipfs-cluster-follow/ipfs-cluster-follow
+    dest: /usr/local/bin/ipfs-cluster-follow
+    mode: "0755"
+    remote_src: true
+  notify:
+    - Restart ipfs-cluster-follow
+
+- name: Generate random peername
+  ansible.builtin.set_fact:
+    cluster_peername: "{{ lookup('password', '/dev/null length=8 chars=hexdigits') }}"
+
+- name: Create ipfs-cluster-follow service
+  ansible.builtin.template:
+    src: ipfs-cluster-follow.service.j2
+    dest: /etc/systemd/system/ipfs-cluster-follow.service
+    mode: "0644"
+  notify:
+    - Restart ipfs-cluster-follow
+
+- name: Create ipfs service
+  ansible.builtin.template:
+    src: ipfs.service.j2
+    dest: /etc/systemd/system/ipfs.service
+    mode: "0644"
+  notify:
+    - Restart ipfs
+
+- name: Check current value of Routing.AcceleratedDHTClient
+  ansible.builtin.command: ipfs config Routing.AcceleratedDHTClient
+  register: ipfs_dht_config
+  changed_when: false # this never changes things, it only gathers data
+
+- name: Enable IPFS Routing.AcceleratedDHTClient
+  ansible.builtin.command: ipfs config --json Routing.AcceleratedDHTClient true
+  notify:
+    - Restart ipfs
+  when: ipfs_dht_config.stdout != "true"
+  changed_when: true
+
+- name: Create IPFS directory
+  ansible.builtin.file:
+    dest: /home/ipfs/.ipfs
+    owner: ipfs
+    group: ipfs
+    state: directory
+    mode: "0755"
+
+- name: Check if IPFS config exists
+  ansible.builtin.stat:
+    path: /home/ipfs/.ipfs/config
+  register: ipfs_config
+
+- name: Initialize IPFS
+  ansible.builtin.command: /usr/local/bin/ipfs init
+  become: true
+  become_user: ipfs
+  args:
+    chdir: /home/ipfs
+  when: not ipfs_config.stat.exists
+  changed_when: true  # Explicitly mark this as a change when it runs
+  notify:
+    - Restart ipfs
+
+## @todo enable once we get gluster working
+# - name: Symlink IPFS blocks directory to gluster brick
+#   ansible.builtin.file:
+#     src: /home/ipfs/.ipfs/blocks
+#     dest: /mnt/g1/.ipfs/blocks
+#     owner: ipfs
+#     group: ipfs
+#     state: link
+#   notify:
+#     - Restart ipfs
diff --git a/ansible/roles/ipfs/templates/ipfs-cluster-follow.service.j2 b/ansible/roles/ipfs/templates/ipfs-cluster-follow.service.j2
new file mode 100644
index 0000000..7b9f6f6
--- /dev/null
+++ b/ansible/roles/ipfs/templates/ipfs-cluster-follow.service.j2
@@ -0,0 +1,14 @@
+[Unit]
+Description=ipfs-cluster-follow
+
+[Service]
+Type=simple
+Environment=CLUSTER_PEERNAME="{{cluster_peername}}" 
+ExecStart=/usr/local/bin/ipfs-cluster-follow futureporn.net run --init https://futureporn.net/api/service.json
+User=ipfs
+Restart=always
+RestartSec=10
+
+[Install]
+WantedBy=multi-user.target
+
diff --git a/ansible/roles/ipfs/templates/ipfs.service.j2 b/ansible/roles/ipfs/templates/ipfs.service.j2
new file mode 100644
index 0000000..a684673
--- /dev/null
+++ b/ansible/roles/ipfs/templates/ipfs.service.j2
@@ -0,0 +1,14 @@
+[Unit]
+Description=IPFS Daemon
+
+[Service]
+Type=simple
+Environment=IPFS_PATH=/home/ipfs/.ipfs
+ExecStart=/usr/local/bin/ipfs daemon
+User=ipfs
+Restart=always
+RestartSec=10
+
+[Install]
+WantedBy=multi-user.target
+
diff --git a/ansible/site.yml b/ansible/site.yml
index b37cdb9..33774c6 100644
--- a/ansible/site.yml
+++ b/ansible/site.yml
@@ -16,6 +16,12 @@
   roles:
     - common
 
+- name: Set up IPFS
+  hosts: ipfs
+  gather_facts: true
+  roles:
+    - ipfs
+
 - name: Assert Capture dependencies
   hosts: capture
   gather_facts: true
@@ -55,3 +61,4 @@
   become: false
   roles:
     - load_balancer
+
diff --git a/apps/bright/assets/js/hooks/uppy_hook.js b/apps/bright/assets/js/hooks/uppy_hook.js
index 30a92e6..c6e74e7 100644
--- a/apps/bright/assets/js/hooks/uppy_hook.js
+++ b/apps/bright/assets/js/hooks/uppy_hook.js
@@ -44,15 +44,16 @@ const UppyHook = {
           })
         },
         createMultipartUpload(file) {
-          // console.log(`createMultipartUpload with file ${JSON.stringify(file)}`)
+          console.log(`createMultipartUpload with file ${JSON.stringify(file)}`)
           let { name, type } = file.meta
-          let payload = { name, type }
+          let size = file.size
+          let payload = { name, type, size }
           return new Promise((resolve) => {
             live.pushEvent('initiate_multipart_upload', payload, (reply, ref) => {
-              // console.log(`payload=${JSON.stringify(payload)}`)
-              // console.log(`initiate_multipart_upload pushEvent response callback.`)
-              // console.log(`got reply=${JSON.stringify(reply)}`)
-              // console.log(`got ref=${JSON.stringify(ref)}`)
+              console.log(`payload=${JSON.stringify(payload)}`)
+              console.log(`initiate_multipart_upload pushEvent response callback.`)
+              console.log(`got reply=${JSON.stringify(reply)}`)
+              console.log(`got ref=${JSON.stringify(ref)}`)
 
               let output = {
                 uploadId: reply?.upload_id,
diff --git a/apps/bright/config/runtime.exs b/apps/bright/config/runtime.exs
index b5cbed2..f1400d9 100644
--- a/apps/bright/config/runtime.exs
+++ b/apps/bright/config/runtime.exs
@@ -49,11 +49,11 @@ config :ex_aws,
   ]
 
 if config_env() == :prod do
-  db_host = System.get_env("DB_HOST") || raise "environment variable DB_HOST is missing."
-  db_user = System.get_env("DB_USER") || raise "environment variable DB_USER is missing."
-  db_pass = System.get_env("DB_PASS") || raise "environment variable DB_PASS is missing."
-  db_port = System.get_env("DB_PORT") || raise "environment variable DB_PORT is missing."
-  db_name = System.get_env("DB_NAME") || raise "environment variable DB_NAME is missing."
+  db_host = System.fetch_env!("DB_HOST")
+  db_user = System.fetch_env!("DB_USER")
+  db_pass = System.fetch_env!("DB_PASS")
+  db_port = System.fetch_env!("DB_PORT")
+  db_name = System.fetch_env!("DB_NAME")
 
   maybe_ipv6 = if System.get_env("ECTO_IPV6") in ~w(true 1), do: [:inet6], else: []
 
diff --git a/apps/bright/config/test.exs b/apps/bright/config/test.exs
index 2c2ae5c..f12c663 100644
--- a/apps/bright/config/test.exs
+++ b/apps/bright/config/test.exs
@@ -9,20 +9,24 @@ config :bcrypt_elixir, :log_rounds, 1
 # to provide built-in test partitioning in CI environment.
 # Run `mix help test` for more information.
 config :bright, Bright.Repo,
-  database: System.get_env("DB_NAME", "bright"),
   # database: "bright_test#{System.get_env("MIX_TEST_PARTITION")}",
-  hostname: System.get_env("DB_HOST", "localhost"),
-  username: System.get_env("DB_USER", "postgres"),
-  password: System.get_env("DB_PASS", "password"),
-  port: System.get_env("DB_PORT", "5433"),
+  database: System.fetch_env!("DB_NAME"),
+  hostname: System.fetch_env!("DB_HOST"),
+  username: System.fetch_env!("DB_USER"),
+  password: System.fetch_env!("DB_PASS"),
+  port: System.fetch_env!("DB_PORT"),
   pool: Ecto.Adapters.SQL.Sandbox,
   pool_size: System.schedulers_online() * 4
 
+IO.puts(
+  "hello this is test.exs and we are checking env vars. db_host=#{System.fetch_env!("DB_HOST")} db_user=#{System.fetch_env!("DB_USER")} db_pass=#{System.fetch_env!("DB_PASS")} db_port=#{System.fetch_env!("DB_PORT")} db_name=#{System.fetch_env!("DB_NAME")}"
+)
+
 # We don't run a server during test. If one is required,
 # you can enable the server option below.
 config :bright, BrightWeb.Endpoint,
   http: [ip: {127, 0, 0, 1}, port: 4002],
-  secret_key_base: "#{System.get_env("SECRET_KEY_BASE")}",
+  secret_key_base: "#{System.fetch_env!("SECRET_KEY_BASE")}",
   server: false
 
 # Prevent Oban from running jobs and plugins during test runs
diff --git a/apps/bright/lib/bright/streams.ex b/apps/bright/lib/bright/streams.ex
index 2b6e32f..fe99cb7 100644
--- a/apps/bright/lib/bright/streams.ex
+++ b/apps/bright/lib/bright/streams.ex
@@ -616,7 +616,9 @@ defmodule Bright.Streams do
 
   """
   def list_uploads do
-    Repo.all(Upload)
+    Upload
+    |> Repo.all()
+    |> Repo.preload([:user])
   end
 
   @doc """
@@ -633,7 +635,11 @@ defmodule Bright.Streams do
       ** (Ecto.NoResultsError)
 
   """
-  def get_upload!(id), do: Repo.get!(Upload, id)
+  def get_upload!(id) do
+    Upload
+    |> Repo.get!(id)
+    |> Repo.preload([:user])
+  end
 
   @doc """
   Creates a upload.
diff --git a/apps/bright/lib/bright/streams/upload.ex b/apps/bright/lib/bright/streams/upload.ex
index 14b7ced..bf8bb79 100644
--- a/apps/bright/lib/bright/streams/upload.ex
+++ b/apps/bright/lib/bright/streams/upload.ex
@@ -6,8 +6,8 @@ defmodule Bright.Streams.Upload do
     field :size, :integer
     field :filename, :string
     field :content_type, :string
-    field :user_id, :id
-    field :vod, :id
+
+    belongs_to :user, Bright.Users.User
 
     timestamps(type: :utc_datetime)
   end
@@ -15,7 +15,8 @@ defmodule Bright.Streams.Upload do
   @doc false
   def changeset(upload, attrs) do
     upload
-    |> cast(attrs, [:filename, :size, :content_type])
-    |> validate_required([:filename, :size, :content_type])
+    |> cast(attrs, [:filename, :size, :content_type, :user_id])
+    |> validate_required([:filename, :size, :content_type, :user_id])
+    |> assoc_constraint(:user)
   end
 end
diff --git a/apps/bright/lib/bright/users/user.ex b/apps/bright/lib/bright/users/user.ex
index 343068b..949d8de 100644
--- a/apps/bright/lib/bright/users/user.ex
+++ b/apps/bright/lib/bright/users/user.ex
@@ -12,6 +12,8 @@ defmodule Bright.Users.User do
     field :patron_tier, :integer
     field :role, :string
 
+    has_many :uploads, Bright.Streams.Upload
+
     timestamps(type: :utc_datetime)
   end
 
@@ -71,3 +73,9 @@ defmodule Bright.Users.User do
     user
   end
 end
+
+defimpl Phoenix.HTML.Safe, for: Bright.Users.User do
+  def to_iodata(user) do
+    Phoenix.HTML.Safe.to_iodata("User #{user.id}")
+  end
+end
diff --git a/apps/bright/lib/bright_web/controllers/upload_html/index.html.heex b/apps/bright/lib/bright_web/controllers/upload_html/index.html.heex
index e329877..3eb0d72 100644
--- a/apps/bright/lib/bright_web/controllers/upload_html/index.html.heex
+++ b/apps/bright/lib/bright_web/controllers/upload_html/index.html.heex
@@ -1,7 +1,7 @@
 <.header>
   Listing Uploads
   <:actions>
-    <.link href={~p"/uploads/new"}>
+    <.link href={~p"/upload"}>
       <.button>New Upload</.button>
     </.link>
   </:actions>
@@ -11,11 +11,11 @@
   <:col :let={upload} label="Filename">{upload.filename}</:col>
   <:col :let={upload} label="Size">{upload.size}</:col>
   <:col :let={upload} label="Content type">{upload.content_type}</:col>
+  <:col :let={upload} label="Uploader">{upload.user}</:col>
   <:action :let={upload}>
     <div class="sr-only">
       <.link navigate={~p"/uploads/#{upload}"}>Show</.link>
     </div>
-    <.link navigate={~p"/uploads/#{upload}/edit"}>Edit</.link>
   </:action>
   <:action :let={upload}>
     <.link href={~p"/uploads/#{upload}"} method="delete" data-confirm="Are you sure?">
diff --git a/apps/bright/lib/bright_web/controllers/upload_html/show.html.heex b/apps/bright/lib/bright_web/controllers/upload_html/show.html.heex
index d015d82..e81c424 100644
--- a/apps/bright/lib/bright_web/controllers/upload_html/show.html.heex
+++ b/apps/bright/lib/bright_web/controllers/upload_html/show.html.heex
@@ -1,17 +1,14 @@
 <.header>
   Upload {@upload.id}
   <:subtitle>This is a upload record from your database.</:subtitle>
-  <:actions>
-    <.link href={~p"/uploads/#{@upload}/edit"}>
-      <.button>Edit upload</.button>
-    </.link>
-  </:actions>
+  <:actions></:actions>
 </.header>
 
 <.list>
   <:item title="Filename">{@upload.filename}</:item>
   <:item title="Size">{@upload.size}</:item>
   <:item title="Content type">{@upload.content_type}</:item>
+  <:item title="Uploader">{@upload.user}</:item>
 </.list>
 
 <.back navigate={~p"/uploads"}>Back to uploads</.back>
diff --git a/apps/bright/lib/bright_web/live/upload_live/index.ex b/apps/bright/lib/bright_web/live/upload_live/index.ex
index 16fda4e..a7a7e3a 100644
--- a/apps/bright/lib/bright_web/live/upload_live/index.ex
+++ b/apps/bright/lib/bright_web/live/upload_live/index.ex
@@ -2,22 +2,12 @@ defmodule BrightWeb.UploadLive.Index do
   use BrightWeb, :live_view
 
   alias Bright.Streams
-  alias Bright.Streams.Upload
   alias Bright.Utils
+
   require Logger
 
+  @impl true
   def mount(_params, _session, socket) do
-    # {:ok,
-    #  socket
-    #  |> assign(:uploaded_files, [])
-    #  |> allow_upload(:vods,
-    #    accept: ~w(.mp4 .mov .ts .avi .mpeg .ogv .webm .3gp .3g2),
-    #    max_entries: 3,
-    #    max_file_size: 80 * 1_000_000_000,
-    #    external: &presign_upload/2
-    #  )}
-
-    # socket = assign(socket, endpoint: System.fetch_env!("UPPY_ENDPOINT"))
     socket =
       socket
       |> assign(:uploaded_files, [])
@@ -28,135 +18,20 @@ defmodule BrightWeb.UploadLive.Index do
   @impl true
   def render(assigns) do
     ~H"""
-    <div>
-
-      <link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/uppy/4.12.2/uppy.min.css" integrity="sha512-oPlr9/HXIlp7YoIRNsexheOu2/P2sEVi8EFQEAWUlHHijx0QbQ9qgihNYmIYtdJP3xOIMbZcnSVhrznIh5DKkg==" crossorigin="anonymous" referrerpolicy="no-referrer" />
-
-
-
-
+    <div class="section">
+      <h2 class="title is-2 mb-3">Upload a VOD</h2>
+      <link
+        rel="stylesheet"
+        href="https://cdnjs.cloudflare.com/ajax/libs/uppy/4.12.2/uppy.min.css"
+        integrity="sha512-oPlr9/HXIlp7YoIRNsexheOu2/P2sEVi8EFQEAWUlHHijx0QbQ9qgihNYmIYtdJP3xOIMbZcnSVhrznIh5DKkg=="
+        crossorigin="anonymous"
+        referrerpolicy="no-referrer"
+      />
       <div id="uppy-dashboard" phx-hook="UppyHook" phx-update="ignore"></div>
-
-
-
     </div>
     """
   end
 
-  # @impl true
-  # def handle_event(
-  #       "presign",
-  #       %{"type" => type, "name" => name},
-  #       socket
-  #     ) do
-  #   {:ok, %{url: url}, socket} = presign_url(name, type, socket)
-  #   Logger.debug("✍️✍️✍️ presign called with name=#{name} type=#{type}. url=#{url}")
-  #   # socket = assign(socket, signatures: {name, url})
-  #   # socket =
-  #   #   socket
-  #   #   |> update(:uploaded_files, &(&1 ++ uploaded_files))
-
-  #   # |> push_event("process_upload", %{name: name, url: url, type: type})
-
-  #   {:reply, %{name: name, type: type, url: url}, socket}
-  # end
-
-  # @impl true
-  # def handle_event("upload_videos", _params, socket) do
-  #   uploaded_files =
-  #     consume_uploaded_entries(socket, :vods, fn %{path: _path}, entry ->
-  #       IO.puts("@todo ⭐⭐⭐ Handling #{entry.client_type} @todo @todo @todo")
-  #       IO.puts(inspect(entry))
-  #     end)
-
-  #   socket =
-  #     socket
-  #     |> update(:uploaded_files, &(&1 ++ uploaded_files))
-
-  #   {:noreply, socket}
-  # end
-
-  # def handle_event("update_preview_srcs", %{"srcs" => srcs}, socket) do
-  #   uploaded_files =
-  #     socket.assigns.uploaded_files
-  #     |> Enum.map(fn entry ->
-  #       if Map.has_key?(srcs, entry.ref) do
-  #         entry
-  #         |> Map.put(:preview_src, Map.fetch!(srcs, entry.ref))
-  #       else
-  #         entry
-  #       end
-  #     end)
-
-  #   socket =
-  #     socket
-  #     |> assign(:uploaded_files, uploaded_files)
-
-  #   {:noreply, socket}
-  # end
-
-  # def handle_event("validate_upload", _params, socket) do
-  #   num_remaining_uploads =
-  #     length(socket.assigns.uploaded_files) - socket.assigns.uploads.vods.max_entries
-
-  #   valid =
-  #     Enum.uniq_by(socket.assigns.uploads.vods.entries, & &1.client_name)
-  #     |> Enum.take(num_remaining_uploads)
-
-  #   socket =
-  #     Enum.reduce(socket.assigns.uploads.vods.entries, socket, fn entry, socket ->
-  #       if entry in valid do
-  #         socket
-  #       else
-  #         socket
-  #         |> cancel_upload(:vods, entry.ref)
-  #         |> put_flash(
-  #           :error,
-  #           "Uploaded files should be unique and cannot exceed #{socket.assigns.uploads.vods.max_entries} total files."
-  #         )
-  #       end
-  #     end)
-
-  #   {:noreply, socket}
-  # end
-
-  # def handle_event("cancel_upload", %{"ref" => ref}, socket) do
-  #   {:noreply, cancel_upload(socket, :vods, ref)}
-  # end
-
-  # def handle_event("cancel_upload", _params, socket) do
-  #   socket =
-  #     Enum.reduce(socket.assigns.uploads.vods.entries, socket, fn entry, socket ->
-  #       cancel_upload(socket, :vods, entry.ref)
-  #     end)
-
-  #   {:noreply, socket}
-  # end
-
-  # def presign_upload(name, type, socket) do
-  #   Logger.debug("presign_upload was called with name=#{inspect(name)} and type=#{inspect(type)}")
-  #   config = ExAws.Config.new(:s3)
-  #   bucket = System.fetch_env!("AWS_BUCKET")
-  #   key = "usc/#{Utils.random_string()}/#{name}"
-
-  #   {:ok, url} =
-  #     ExAws.S3.presigned_url(config, :put, bucket, key,
-  #       expires_in: 3600,
-  #       query_params: [{"Content-Type", type}]
-  #     )
-
-  #   {:ok, %{uploader: "S3", key: key, url: url}, socket}
-  # end
-
-  # # @doc @see https://hexdocs.pm/ex_aws_s3/ExAws.S3.html#presigned_post/4
-  # def presigned_post(name, socket) do
-  #   Logger.debug("presigned_post with name=#{inspect(name)}")
-  #   config = ExAws.Config.new(:s3)
-  #   bucket = System.fetch_env!("AWS_BUCKET")
-  #   key = "usc/#{Utils.random_string()}/#{name}"
-  #   {:ok, url} = ExAws.S3.presigned_post(config, )
-  # end
-
   defp get_s3_config(name) do
     config = ExAws.Config.new(:s3)
     bucket = System.fetch_env!("AWS_BUCKET")
@@ -164,22 +39,15 @@ defmodule BrightWeb.UploadLive.Index do
     %{config: config, bucket: bucket, key: key}
   end
 
+  @impl true
   def handle_event(
         "list_parts",
         %{"upload_id" => upload_id, "key" => key},
         socket
       ) do
-    config = ExAws.Config.new(:s3)
     bucket = System.fetch_env!("AWS_BUCKET")
-    # key = "usc/#{Utils.random_string()}/#{name}"
-    # %{config: config, bucket: bucket, key: key} = get_s3_config(name)
-    case ExAws.S3.list_parts(bucket, key, upload_id) do
-      # <Part>
-      # <ETag>"85f30635602dc09bd85957a6e82a2c21"</ETag>
-      # <LastModified>2023-08-31T18:54:55.693Z</LastModified>
-      # <PartNumber>1</PartNumber>
-      # <Size>11</Size>
 
+    case ExAws.S3.list_parts(bucket, key, upload_id) do
       {:ok, part: %{etag: etag, partNumber: partNumber, size: size}} ->
         Logger.debug("🦠🦠🦠 we got an etag from list_parts. etag=#{inspect(etag)}")
         {:reply, %{etag: etag, partNumber: partNumber, size: size}, socket}
@@ -191,9 +59,10 @@ defmodule BrightWeb.UploadLive.Index do
   end
 
   # @doc @see https://hexdocs.pm/ex_aws_s3/ExAws.S3.html#initiate_multipart_upload/3
+  @impl true
   def handle_event(
         "initiate_multipart_upload",
-        %{"name" => name, "type" => type},
+        %{"name" => name, "type" => type, "size" => size},
         socket
       ) do
     %{config: config, bucket: bucket, key: key} = get_s3_config(name)
@@ -203,9 +72,22 @@ defmodule BrightWeb.UploadLive.Index do
     case ExAws.request(operation, config) do
       {:ok, %{body: %{key: key, upload_id: upload_id}} = response} ->
         Logger.debug(
-          "Multipart upload initiated. Upload ID: #{upload_id}, Key: #{key}, response=#{inspect(response)}"
+          "👤 Multipart upload initiated. Upload ID: #{upload_id}, Key: #{key}, response=#{inspect(response)}, user.id=#{socket.assigns.current_user.id}"
         )
 
+        case Streams.create_upload(%{
+               filename: name,
+               content_type: type,
+               user_id: socket.assigns.current_user.id,
+               size: size
+             }) do
+          {:ok, stream} ->
+            Logger.debug("created stream #{stream.id}")
+
+          {:error, reason} ->
+            Logger.error("failed to create stream. reason=#{inspect(reason)}")
+        end
+
         {:reply, %{upload_id: upload_id, key: key}, socket}
 
       {:error, reason} ->
@@ -214,6 +96,7 @@ defmodule BrightWeb.UploadLive.Index do
     end
   end
 
+  @impl true
   def handle_event(
         "abort_multipart_upload",
         %{"key" => key, "uploadId" => upload_id},
@@ -235,6 +118,7 @@ defmodule BrightWeb.UploadLive.Index do
     end
   end
 
+  @impl true
   def handle_event(
         "get_upload_parameters",
         %{"type" => type, "name" => name},
@@ -253,6 +137,7 @@ defmodule BrightWeb.UploadLive.Index do
     {:reply, %{type: type, method: "PUT", url: url}, socket}
   end
 
+  @impl true
   def handle_event(
         "complete_multipart_upload",
         %{"key" => key, "uploadId" => upload_id, "parts" => parts},
@@ -284,6 +169,7 @@ defmodule BrightWeb.UploadLive.Index do
     end
   end
 
+  @impl true
   def handle_event(
         "sign_part",
         %{"body" => _body, "key" => key, "partNumber" => part_number, "uploadId" => upload_id},
@@ -308,24 +194,6 @@ defmodule BrightWeb.UploadLive.Index do
     {:reply, %{url: url}, socket}
   end
 
-  defp join_refs(entries), do: Enum.join(entries, ",")
   def error_to_string(:too_large), do: "File too large!"
   def error_to_string(:not_accepted), do: "Bad file type!"
-
-  defp to_megabytes_or_kilobytes(bytes) when is_integer(bytes) do
-    case bytes do
-      b when b < 1_048_576 ->
-        kilobytes = (b / 1024) |> Float.round(1)
-
-        if kilobytes < 1 do
-          "#{kilobytes}MB"
-        else
-          "#{round(kilobytes)}KB"
-        end
-
-      _ ->
-        megabytes = (bytes / 1_048_576) |> Float.round(1)
-        "#{megabytes}MB"
-    end
-  end
 end
diff --git a/apps/bright/lib/bright_web/live/upload_live/show.html.heex b/apps/bright/lib/bright_web/live/upload_live/show.html.heex
index b861be5..cf8fea3 100644
--- a/apps/bright/lib/bright_web/live/upload_live/show.html.heex
+++ b/apps/bright/lib/bright_web/live/upload_live/show.html.heex
@@ -1,11 +1,7 @@
 <.header>
   Upload {@upload.id}
   <:subtitle>This is a upload record from your database.</:subtitle>
-  <:actions>
-    <.link patch={~p"/uploads/#{@upload}/show/edit"} phx-click={JS.push_focus()}>
-      <.button>Edit upload</.button>
-    </.link>
-  </:actions>
+  <:actions></:actions>
 </.header>
 
 <.list>
@@ -15,8 +11,12 @@
 </.list>
 
 <.back navigate={~p"/uploads"}>Back to uploads</.back>
-
-<.modal :if={@live_action == :edit} id="upload-modal" show on_cancel={JS.patch(~p"/uploads/#{@upload}")}>
+<.modal
+  :if={@live_action == :edit}
+  id="upload-modal"
+  show
+  on_cancel={JS.patch(~p"/uploads/#{@upload}")}
+>
   <.live_component
     module={BrightWeb.UploadLive.FormComponent}
     id={@upload.id}
diff --git a/apps/bright/lib/bright_web/router.ex b/apps/bright/lib/bright_web/router.ex
index 83e69e4..83aef6f 100644
--- a/apps/bright/lib/bright_web/router.ex
+++ b/apps/bright/lib/bright_web/router.ex
@@ -2,6 +2,7 @@ defmodule BrightWeb.Router do
   use BrightWeb, :router
 
   import Oban.Web.Router
+  import Redirect
 
   import BrightWeb.AuthController,
     only: [
@@ -75,7 +76,6 @@ defmodule BrightWeb.Router do
     #   get "/vtubers/:id/edit", VtuberController, :edit
     # end
 
-    resources("/uploads", UploadController, only: [:show, :index, :delete])
     resources("/vods", VodController, only: [:create, :new, :edit, :update, :delete])
     resources("/vtubers", VtuberController, only: [:delete])
 
@@ -128,9 +128,10 @@ defmodule BrightWeb.Router do
     live_session :authenticated,
       on_mount: [{BrightWeb.AuthController, :ensure_authenticated}] do
       live("/profile", ProfileLive)
-      live("/upload", UploadLive.Index, :index)
-      # live("/upload/presign", , :)
+      live("/uploads/new", UploadLive.Index, :index)
     end
+
+    resources("/uploads", UploadController, only: [:show, :index, :delete])
   end
 
   scope "/feeds", BrightWeb do
@@ -144,6 +145,8 @@ defmodule BrightWeb.Router do
     get("/health", PageController, :health)
   end
 
+  redirect("/upload", "/uploads/new", :permanent, preserve_query_string: true)
+
   # Enable LiveDashboard and Swoosh mailbox preview in development
   if Application.compile_env(:bright, :dev_routes) do
     # If you want to use the LiveDashboard in production, you should put
diff --git a/apps/bright/mix.exs b/apps/bright/mix.exs
index f382cd0..d803366 100644
--- a/apps/bright/mix.exs
+++ b/apps/bright/mix.exs
@@ -37,6 +37,7 @@ defmodule Bright.MixProject do
   # Type `mix help deps` for examples and options.
   defp deps do
     [
+      {:redirect, "~> 0.4.0"},
       {:bcrypt_elixir, "~> 3.0"},
       {:phoenix, "~> 1.7.17"},
       {:phoenix_ecto, "~> 4.5"},
diff --git a/apps/bright/priv/repo/migrations/20250323033920_create_uploads.exs b/apps/bright/priv/repo/migrations/20250323033920_create_uploads.exs
deleted file mode 100644
index 0f5271c..0000000
--- a/apps/bright/priv/repo/migrations/20250323033920_create_uploads.exs
+++ /dev/null
@@ -1,18 +0,0 @@
-defmodule Bright.Repo.Migrations.CreateUploads do
-  use Ecto.Migration
-
-  def change do
-    create table(:uploads) do
-      add :filename, :string
-      add :size, :integer
-      add :content_type, :string
-      add :user_id, references(:users, on_delete: :nothing)
-      add :vod, references(:vods, on_delete: :nothing)
-
-      timestamps(type: :utc_datetime)
-    end
-
-    create index(:uploads, [:user_id])
-    create index(:uploads, [:vod])
-  end
-end
diff --git a/apps/bright/test/bright/platforms_test.exs b/apps/bright/test/bright/platforms_test.exs
index 3038fc2..fe803e5 100644
--- a/apps/bright/test/bright/platforms_test.exs
+++ b/apps/bright/test/bright/platforms_test.exs
@@ -104,17 +104,20 @@ defmodule Bright.PlatformsTest do
     @invalid_attrs %{url: nil}
 
     test "list_platform_aliases/0 returns all platform_aliases" do
-      platform_alias = platform_alias_fixture()
+      platform = platform_fixture()
+      platform_alias = platform_alias_fixture(%{platform_id: platform.id}).preload([:platform])
       assert Platforms.list_platform_aliases() == [platform_alias]
     end
 
     test "get_platform_alias!/1 returns the platform_alias with given id" do
-      platform_alias = platform_alias_fixture()
+      platform = platform_fixture()
+      platform_alias = platform_alias_fixture(%{platform_id: platform.id})
       assert Platforms.get_platform_alias!(platform_alias.id) == platform_alias
     end
 
     test "create_platform_alias/1 with valid data creates a platform_alias" do
-      valid_attrs = %{url: "some url"}
+      platform = platform_fixture()
+      valid_attrs = %{url: "some url", platform_id: platform.id}
 
       assert {:ok, %PlatformAlias{} = platform_alias} =
                Platforms.create_platform_alias(valid_attrs)
@@ -127,7 +130,8 @@ defmodule Bright.PlatformsTest do
     end
 
     test "update_platform_alias/2 with valid data updates the platform_alias" do
-      platform_alias = platform_alias_fixture()
+      platform = platform_fixture()
+      platform_alias = platform_alias_fixture(%{platform_id: platform.id})
       update_attrs = %{url: "some updated url"}
 
       assert {:ok, %PlatformAlias{} = platform_alias} =
@@ -137,7 +141,8 @@ defmodule Bright.PlatformsTest do
     end
 
     test "update_platform_alias/2 with invalid data returns error changeset" do
-      platform_alias = platform_alias_fixture()
+      platform = platform_fixture()
+      platform_alias = platform_alias_fixture(%{platform_id: platform.id})
 
       assert {:error, %Ecto.Changeset{}} =
                Platforms.update_platform_alias(platform_alias, @invalid_attrs)
@@ -152,7 +157,8 @@ defmodule Bright.PlatformsTest do
     end
 
     test "change_platform_alias/1 returns a platform_alias changeset" do
-      platform_alias = platform_alias_fixture()
+      platform = platform_fixture()
+      platform_alias = platform_alias_fixture(%{platform_id: platform.id})
       assert %Ecto.Changeset{} = Platforms.change_platform_alias(platform_alias)
     end
   end
diff --git a/apps/bright/test/bright/streams_test.exs b/apps/bright/test/bright/streams_test.exs
index 14277e5..2e9875f 100644
--- a/apps/bright/test/bright/streams_test.exs
+++ b/apps/bright/test/bright/streams_test.exs
@@ -188,62 +188,4 @@ defmodule Bright.StreamsTest do
       # assert_received {:progress, %{stage: :generating_thumbnail, done: 1, total: 1}}
     end
   end
-
-  describe "uploads" do
-    alias Bright.Streams.Upload
-
-    import Bright.StreamsFixtures
-
-    @invalid_attrs %{size: nil, filename: nil, content_type: nil}
-
-    test "list_uploads/0 returns all uploads" do
-      upload = upload_fixture()
-      assert Streams.list_uploads() == [upload]
-    end
-
-    test "get_upload!/1 returns the upload with given id" do
-      upload = upload_fixture()
-      assert Streams.get_upload!(upload.id) == upload
-    end
-
-    test "create_upload/1 with valid data creates a upload" do
-      valid_attrs = %{size: 42, filename: "some filename", content_type: "some content_type"}
-
-      assert {:ok, %Upload{} = upload} = Streams.create_upload(valid_attrs)
-      assert upload.size == 42
-      assert upload.filename == "some filename"
-      assert upload.content_type == "some content_type"
-    end
-
-    test "create_upload/1 with invalid data returns error changeset" do
-      assert {:error, %Ecto.Changeset{}} = Streams.create_upload(@invalid_attrs)
-    end
-
-    test "update_upload/2 with valid data updates the upload" do
-      upload = upload_fixture()
-      update_attrs = %{size: 43, filename: "some updated filename", content_type: "some updated content_type"}
-
-      assert {:ok, %Upload{} = upload} = Streams.update_upload(upload, update_attrs)
-      assert upload.size == 43
-      assert upload.filename == "some updated filename"
-      assert upload.content_type == "some updated content_type"
-    end
-
-    test "update_upload/2 with invalid data returns error changeset" do
-      upload = upload_fixture()
-      assert {:error, %Ecto.Changeset{}} = Streams.update_upload(upload, @invalid_attrs)
-      assert upload == Streams.get_upload!(upload.id)
-    end
-
-    test "delete_upload/1 deletes the upload" do
-      upload = upload_fixture()
-      assert {:ok, %Upload{}} = Streams.delete_upload(upload)
-      assert_raise Ecto.NoResultsError, fn -> Streams.get_upload!(upload.id) end
-    end
-
-    test "change_upload/1 returns a upload changeset" do
-      upload = upload_fixture()
-      assert %Ecto.Changeset{} = Streams.change_upload(upload)
-    end
-  end
 end
diff --git a/apps/bright/test/bright_web/controllers/platform_alias_controller_test.exs b/apps/bright/test/bright_web/controllers/platform_alias_controller_test.exs
index 40c8655..456873c 100644
--- a/apps/bright/test/bright_web/controllers/platform_alias_controller_test.exs
+++ b/apps/bright/test/bright_web/controllers/platform_alias_controller_test.exs
@@ -8,6 +8,8 @@ defmodule BrightWeb.PlatformAliasControllerTest do
   @invalid_attrs %{url: nil}
 
   describe "index" do
+    setup [:create_admin]
+
     test "lists all platform_aliases", %{conn: conn} do
       conn = get(conn, ~p"/platform_aliases")
       assert html_response(conn, 200) =~ "Listing Platform aliases"
@@ -15,6 +17,8 @@ defmodule BrightWeb.PlatformAliasControllerTest do
   end
 
   describe "new platform_alias" do
+    setup [:create_admin]
+
     test "renders form", %{conn: conn} do
       conn = get(conn, ~p"/platform_aliases/new")
       assert html_response(conn, 200) =~ "New Platform alias"
@@ -22,9 +26,20 @@ defmodule BrightWeb.PlatformAliasControllerTest do
   end
 
   describe "create platform_alias" do
-    test "redirects to show when data is valid", %{conn: conn} do
-      conn = post(conn, ~p"/platform_aliases", platform_alias: @create_attrs)
+    setup [:create_admin]
 
+    test "redirects to show when data is valid", %{conn: conn} do
+      platform = platform_fixture()
+
+      create_attrs = %{
+        url: "https://example.com",
+        platform_id: platform.id
+      }
+
+      conn =
+        post(conn, ~p"/platform_aliases", platform_alias: create_attrs)
+
+      # IO.puts(conn.resp_body)
       assert %{id: id} = redirected_params(conn)
       assert redirected_to(conn) == ~p"/platform_aliases/#{id}"
 
@@ -39,7 +54,7 @@ defmodule BrightWeb.PlatformAliasControllerTest do
   end
 
   describe "edit platform_alias" do
-    setup [:create_platform_alias]
+    setup [:create_platform_alias, :create_admin]
 
     test "renders form for editing chosen platform_alias", %{
       conn: conn,
@@ -51,7 +66,7 @@ defmodule BrightWeb.PlatformAliasControllerTest do
   end
 
   describe "update platform_alias" do
-    setup [:create_platform_alias]
+    setup [:create_platform_alias, :create_admin]
 
     test "redirects when data is valid", %{conn: conn, platform_alias: platform_alias} do
       conn = put(conn, ~p"/platform_aliases/#{platform_alias}", platform_alias: @update_attrs)
@@ -68,7 +83,7 @@ defmodule BrightWeb.PlatformAliasControllerTest do
   end
 
   describe "delete platform_alias" do
-    setup [:create_platform_alias]
+    setup [:create_platform_alias, :create_admin]
 
     test "deletes chosen platform_alias", %{conn: conn, platform_alias: platform_alias} do
       conn = delete(conn, ~p"/platform_aliases/#{platform_alias}")
@@ -81,7 +96,8 @@ defmodule BrightWeb.PlatformAliasControllerTest do
   end
 
   defp create_platform_alias(_) do
-    platform_alias = platform_alias_fixture()
+    platform = platform_fixture()
+    platform_alias = platform_alias_fixture(%{platform_id: platform.id})
     %{platform_alias: platform_alias}
   end
 end
diff --git a/apps/bright/test/bright_web/controllers/upload_controller_test.exs b/apps/bright/test/bright_web/controllers/upload_controller_test.exs
index d61ef8b..d7ddf26 100644
--- a/apps/bright/test/bright_web/controllers/upload_controller_test.exs
+++ b/apps/bright/test/bright_web/controllers/upload_controller_test.exs
@@ -3,9 +3,14 @@ defmodule BrightWeb.UploadControllerTest do
 
   import Bright.StreamsFixtures
 
-  @create_attrs %{size: 42, filename: "some filename", content_type: "some content_type"}
-  @update_attrs %{size: 43, filename: "some updated filename", content_type: "some updated content_type"}
-  @invalid_attrs %{size: nil, filename: nil, content_type: nil}
+  @create_attrs %{
+    size: 42,
+    filename: "some filename",
+    content_type: "some content_type",
+    user_id: 1
+  }
+
+  @invalid_attrs %{size: nil, filename: nil, content_type: nil, user_id: nil}
 
   describe "index" do
     test "lists all uploads", %{conn: conn} do
@@ -14,56 +19,6 @@ defmodule BrightWeb.UploadControllerTest do
     end
   end
 
-  describe "new upload" do
-    test "renders form", %{conn: conn} do
-      conn = get(conn, ~p"/uploads/new")
-      assert html_response(conn, 200) =~ "New Upload"
-    end
-  end
-
-  describe "create upload" do
-    test "redirects to show when data is valid", %{conn: conn} do
-      conn = post(conn, ~p"/uploads", upload: @create_attrs)
-
-      assert %{id: id} = redirected_params(conn)
-      assert redirected_to(conn) == ~p"/uploads/#{id}"
-
-      conn = get(conn, ~p"/uploads/#{id}")
-      assert html_response(conn, 200) =~ "Upload #{id}"
-    end
-
-    test "renders errors when data is invalid", %{conn: conn} do
-      conn = post(conn, ~p"/uploads", upload: @invalid_attrs)
-      assert html_response(conn, 200) =~ "New Upload"
-    end
-  end
-
-  describe "edit upload" do
-    setup [:create_upload]
-
-    test "renders form for editing chosen upload", %{conn: conn, upload: upload} do
-      conn = get(conn, ~p"/uploads/#{upload}/edit")
-      assert html_response(conn, 200) =~ "Edit Upload"
-    end
-  end
-
-  describe "update upload" do
-    setup [:create_upload]
-
-    test "redirects when data is valid", %{conn: conn, upload: upload} do
-      conn = put(conn, ~p"/uploads/#{upload}", upload: @update_attrs)
-      assert redirected_to(conn) == ~p"/uploads/#{upload}"
-
-      conn = get(conn, ~p"/uploads/#{upload}")
-      assert html_response(conn, 200) =~ "some updated filename"
-    end
-
-    test "renders errors when data is invalid", %{conn: conn, upload: upload} do
-      conn = put(conn, ~p"/uploads/#{upload}", upload: @invalid_attrs)
-      assert html_response(conn, 200) =~ "Edit Upload"
-    end
-  end
-
   describe "delete upload" do
     setup [:create_upload]
 
diff --git a/apps/bright/test/bright_web/live/upload_live_test.exs b/apps/bright/test/bright_web/live/upload_live_test.exs
index 3e789d8..b414e55 100644
--- a/apps/bright/test/bright_web/live/upload_live_test.exs
+++ b/apps/bright/test/bright_web/live/upload_live_test.exs
@@ -1,113 +1,117 @@
-defmodule BrightWeb.UploadLiveTest do
-  use BrightWeb.ConnCase
+# defmodule BrightWeb.UploadLiveTest do
+#   use BrightWeb.ConnCase
 
-  import Phoenix.LiveViewTest
-  import Bright.StreamsFixtures
+#   import Phoenix.LiveViewTest
+#   import Bright.StreamsFixtures
 
-  @create_attrs %{size: 42, filename: "some filename", content_type: "some content_type"}
-  @update_attrs %{size: 43, filename: "some updated filename", content_type: "some updated content_type"}
-  @invalid_attrs %{size: nil, filename: nil, content_type: nil}
+#   @create_attrs %{size: 42, filename: "some filename", content_type: "some content_type"}
+#   @update_attrs %{
+#     size: 43,
+#     filename: "some updated filename",
+#     content_type: "some updated content_type"
+#   }
+#   @invalid_attrs %{size: nil, filename: nil, content_type: nil}
 
-  defp create_upload(_) do
-    upload = upload_fixture()
-    %{upload: upload}
-  end
+#   defp create_upload(_) do
+#     upload = upload_fixture()
+#     %{upload: upload}
+#   end
 
-  describe "Index" do
-    setup [:create_upload]
+#   describe "Index" do
+#     setup [:create_upload, :create_user]
 
-    test "lists all uploads", %{conn: conn, upload: upload} do
-      {:ok, _index_live, html} = live(conn, ~p"/uploads")
+#     test "lists all uploads", %{conn: conn, upload: upload} do
+#       {:ok, _index_live, html} = live(conn, ~p"/uploads")
 
-      assert html =~ "Listing Uploads"
-      assert html =~ upload.filename
-    end
+#       assert html =~ "Listing Uploads"
+#       assert html =~ upload.filename
+#     end
 
-    test "saves new upload", %{conn: conn} do
-      {:ok, index_live, _html} = live(conn, ~p"/uploads")
+#     test "saves new upload", %{conn: conn} do
+#       {:ok, index_live, _html} = live(conn, ~p"/uploads")
 
-      assert index_live |> element("a", "New Upload") |> render_click() =~
-               "New Upload"
+#       assert index_live |> element("a", "New Upload") |> render_click() =~
+#                "New Upload"
 
-      assert_patch(index_live, ~p"/uploads/new")
+#       assert_patch(index_live, ~p"/uploads/new")
 
-      assert index_live
-             |> form("#upload-form", upload: @invalid_attrs)
-             |> render_change() =~ "can&#39;t be blank"
+#       assert index_live
+#              |> form("#upload-form", upload: @invalid_attrs)
+#              |> render_change() =~ "can&#39;t be blank"
 
-      assert index_live
-             |> form("#upload-form", upload: @create_attrs)
-             |> render_submit()
+#       assert index_live
+#              |> form("#upload-form", upload: @create_attrs)
+#              |> render_submit()
 
-      assert_patch(index_live, ~p"/uploads")
+#       assert_patch(index_live, ~p"/uploads")
 
-      html = render(index_live)
-      assert html =~ "Upload created successfully"
-      assert html =~ "some filename"
-    end
+#       html = render(index_live)
+#       assert html =~ "Upload created successfully"
+#       assert html =~ "some filename"
+#     end
 
-    test "updates upload in listing", %{conn: conn, upload: upload} do
-      {:ok, index_live, _html} = live(conn, ~p"/uploads")
+#     test "updates upload in listing", %{conn: conn, upload: upload} do
+#       {:ok, index_live, _html} = live(conn, ~p"/uploads")
 
-      assert index_live |> element("#uploads-#{upload.id} a", "Edit") |> render_click() =~
-               "Edit Upload"
+#       assert index_live |> element("#uploads-#{upload.id} a", "Edit") |> render_click() =~
+#                "Edit Upload"
 
-      assert_patch(index_live, ~p"/uploads/#{upload}/edit")
+#       assert_patch(index_live, ~p"/uploads/#{upload}/edit")
 
-      assert index_live
-             |> form("#upload-form", upload: @invalid_attrs)
-             |> render_change() =~ "can&#39;t be blank"
+#       assert index_live
+#              |> form("#upload-form", upload: @invalid_attrs)
+#              |> render_change() =~ "can&#39;t be blank"
 
-      assert index_live
-             |> form("#upload-form", upload: @update_attrs)
-             |> render_submit()
+#       assert index_live
+#              |> form("#upload-form", upload: @update_attrs)
+#              |> render_submit()
 
-      assert_patch(index_live, ~p"/uploads")
+#       assert_patch(index_live, ~p"/uploads")
 
-      html = render(index_live)
-      assert html =~ "Upload updated successfully"
-      assert html =~ "some updated filename"
-    end
+#       html = render(index_live)
+#       assert html =~ "Upload updated successfully"
+#       assert html =~ "some updated filename"
+#     end
 
-    test "deletes upload in listing", %{conn: conn, upload: upload} do
-      {:ok, index_live, _html} = live(conn, ~p"/uploads")
+#     test "deletes upload in listing", %{conn: conn, upload: upload} do
+#       {:ok, index_live, _html} = live(conn, ~p"/uploads")
 
-      assert index_live |> element("#uploads-#{upload.id} a", "Delete") |> render_click()
-      refute has_element?(index_live, "#uploads-#{upload.id}")
-    end
-  end
+#       assert index_live |> element("#uploads-#{upload.id} a", "Delete") |> render_click()
+#       refute has_element?(index_live, "#uploads-#{upload.id}")
+#     end
+#   end
 
-  describe "Show" do
-    setup [:create_upload]
+#   describe "Show" do
+#     setup [:create_upload, :create_user]
 
-    test "displays upload", %{conn: conn, upload: upload} do
-      {:ok, _show_live, html} = live(conn, ~p"/uploads/#{upload}")
+#     test "displays upload", %{conn: conn, upload: upload} do
+#       {:ok, _show_live, html} = live(conn, ~p"/uploads/#{upload}")
 
-      assert html =~ "Show Upload"
-      assert html =~ upload.filename
-    end
+#       assert html =~ "Show Upload"
+#       assert html =~ upload.filename
+#     end
 
-    test "updates upload within modal", %{conn: conn, upload: upload} do
-      {:ok, show_live, _html} = live(conn, ~p"/uploads/#{upload}")
+#     test "updates upload within modal", %{conn: conn, upload: upload} do
+#       {:ok, show_live, _html} = live(conn, ~p"/uploads/#{upload}")
 
-      assert show_live |> element("a", "Edit") |> render_click() =~
-               "Edit Upload"
+#       assert show_live |> element("a", "Edit") |> render_click() =~
+#                "Edit Upload"
 
-      assert_patch(show_live, ~p"/uploads/#{upload}/show/edit")
+#       assert_patch(show_live, ~p"/uploads/#{upload}/show/edit")
 
-      assert show_live
-             |> form("#upload-form", upload: @invalid_attrs)
-             |> render_change() =~ "can&#39;t be blank"
+#       assert show_live
+#              |> form("#upload-form", upload: @invalid_attrs)
+#              |> render_change() =~ "can&#39;t be blank"
 
-      assert show_live
-             |> form("#upload-form", upload: @update_attrs)
-             |> render_submit()
+#       assert show_live
+#              |> form("#upload-form", upload: @update_attrs)
+#              |> render_submit()
 
-      assert_patch(show_live, ~p"/uploads/#{upload}")
+#       assert_patch(show_live, ~p"/uploads/#{upload}")
 
-      html = render(show_live)
-      assert html =~ "Upload updated successfully"
-      assert html =~ "some updated filename"
-    end
-  end
-end
+#       html = render(show_live)
+#       assert html =~ "Upload updated successfully"
+#       assert html =~ "some updated filename"
+#     end
+#   end
+# end
diff --git a/devbox.json b/devbox.json
index d11bf5b..10813f3 100644
--- a/devbox.json
+++ b/devbox.json
@@ -31,8 +31,8 @@
       "tunnel": "dotenvx run -f ./.env.development -- chisel client bright.fp.sbtp.xyz:9090 R:4000",
       "backup": "docker exec -t postgres_db pg_dumpall -c -U postgres > ./backups/dev_`date +%Y-%m-%d_%H_%M_%S`.sql",
       "act": "dotenvx run -f ./.env.testing -- act -W ./.gitea/workflows --secret-file .env.development",
-      "act:builder": "dotenvx run -f ./.env.testing -- act -W ./.gitea/workflows/builder.yaml --secret-file .env.testing --var-file .env.testing --insecure-secrets",
-      "act:tests": "dotenvx run -f ./.env.testing -- act -W ./.gitea/workflows/tests.yaml --secret-file .env.testing --var-file .env.testing --insecure-secrets",
+      "act:builder": "dotenvx run -f ./.env.testing -- act --env-file .env.testing -W ./.gitea/workflows/builder.yaml --secret-file .env.testing --var-file .env.testing --insecure-secrets",
+      "act:tests": "dotenvx run -f ./.env.testing -- act --env-file .env.testing -W ./.gitea/workflows/tests.yaml --secret-file .env.testing --var-file .env.testing --insecure-secrets",
       "bright:compile:watch": "cd ./apps/bright && find . -type f -name \"*.ex\" -o -name \"*.exs\" | entr -r mix compile --warnings-as-errors",
       "bright:compile:watch2": "cd ./apps/bright && pnpx chokidar-cli \"**/*\" -i \"deps/**\" -i \"_build/**\" -c \"mix compile --warnings-as-errors\"",
       "bright:dev": "cd ./apps/bright && dotenvx run -f ../../.env.development -e MIX_ENV=dev -- mix phx.server",
diff --git a/terraform/main.tf b/terraform/main.tf
index 82ae4a1..fca5a78 100644
--- a/terraform/main.tf
+++ b/terraform/main.tf
@@ -3,7 +3,13 @@
 // @see https://grep.app/search?q=for+tuple+in+regexall%28
 // @see https://github.com/lrmendess/open-source-datalake/blob/main/minio.tf
 locals {
-  envs = { for tuple in regexall("(.*)=(.*)", file("../.env")) : tuple[0] => sensitive(tuple[1]) }
+  envs = { for tuple in regexall("(.*)=(.*)", file("../.env.production")) : tuple[0] => sensitive(tuple[1]) }
+}
+
+variable "ipfs_hosts" {
+  description = "List of IP addresses for IPFS nodes"
+  type        = list(string)
+  default     = ["161.97.186.203", "38.242.193.246"]
 }
 
 
@@ -216,6 +222,20 @@ resource "vultr_instance" "tracker" {
   reserved_ip_id  = vultr_reserved_ip.futureporn_tracker_ip.id
 }
 
+resource "ansible_host" "ipfs_vps" {
+  for_each = { for idx, host in var.ipfs_hosts : idx => host }
+  name     = each.value
+  groups   = ["ipfs"]
+
+  variables = {
+    ansible_user = "root"
+    ansible_host = each.value
+  }
+}
+
+
+
+
 resource "ansible_host" "capture_vps" {
   for_each = { for idx, host in vultr_instance.capture_vps : idx => host }
   name     = each.value.hostname