The Utah Organoids DataJoint pipelines facilitate cerebral organoid characterization and electrophysiology (ephys) data analysis.
-
Organoid Generation Pipeline: Manages metadata for organoid generation protocols, tracking the process from induced pluripotent stem cells (iPSCs) to single neural rosettes (SNRs) to mature organoids.
-
Array Ephys Pipeline: Manages and analyzes ephys recordings, including spike sorting and quality metrics.
- Request Access: Contact the DataJoint support team for an account.
- Log in: Use your DataJoint credentials to access:
- works.datajoint.com (run notebooks & manage computations)
- Organoids SciViz (enter experimental metadata)
- Database connections (access data through the pipeline)
- Log into works.datajoint.com and navigate to the
Notebooktab. - Run EXPLORE_pipeline_architecture.ipynb to visualize the data pipeline structure, including key schemas, tables, and their relationships.
- Log into Organoids SciViz with your DataJoint credentials (username and password).
- Enter data in the corresponding sections:
Userpage → if you are a new experimenter, register a new experimenter.Lineagepage → create new “Lineage” and “Sequence” and submit.Stem Cellpage → register new “Stem Cell” data.Inductionpage → add new “Induction Culture” and “Induction Culture Condition”Post Inductionpage → add new “Post Induction Culture” and “Post Induction Culture Condition”Isolated Rosettepage → add new “Isolated Rosette Culture” and “Isolated Rosette Culture Condition”Organoidpage → add new “Organoid Culture” and “Organoid Culture Condition”Experimentpage → log new experiments performed on a particular organoid- Include metadata: organoid ID, datetime, experimenter, condition, etc.
- Provide the experiment data directory — the relative path to where the acquired data is stored.
- Ensure data follows the file structure guidelines.
- Request Axon credentials from the DataJoint support team.
- Set up your local machine (if you haven't already):
- Install the pipeline code.
- Configure axon settings (Cloud upload configuration).
- Upload data via the cloud upload notebook using either:
- Jupyter Notebook Server:
- Open a terminal or command prompt.
- Activate the
utah_organoidsenvironment withconda activate utah_organoids. - Ensure
Jupyteris installed in theutah_organoidsenvironment. If not, install it by runningconda install jupyter. - Navigate to the
utah_organoids/notebooksdirectory in the terminal. - Run
jupyter notebookin the terminal which will open the Jupyter notebook web interface. - Click on the notebook there (
UPLOAD_session_data_to_cloud.ipynb) and follow the instructions to upload your data to the cloud. - Note: to execute each code cell sequentially, press
Shift + Enteron your keyboard or click "Run". - Close the browser tab and stop Jupyter with
Ctrl + Cin the terminal when you are done with the upload and notebook.
- Visual Studio Code (VS Code):
- Install VS Code and the Python extension.
- Select the kernel for the notebook by clicking on the kernel name
utah_organoidsin the top right corner of the notebook. - Open the
CREATE_new_session_with_cloud_upload.ipynbnotebook in VS Code. - Click on the "Run Cell" button in the top right corner of each code cell to execute the code.
- Follow the instructions in the notebook to upload your data to the cloud.
- Jupyter Notebook Server:
- Navigate to works.datajoint.com and open the
Dashboardtab. - Click on
Plots>MUA Trace Plots, then select a data entry to explore the MUA results. The interactive plot allows you to zoom in and out of the raw traces and examine detected peaks. - (Optional) For a more detailed analysis, go to the
Notebooktab on works.datajoint.com and run the EXPLORE_MUA_analysis.ipynb notebook to inspect theMUAschema in depth.
- Log into works.datajoint.com and navigate to the
Notebooktab. - Open and execute CREATE_new_session.ipynb.
- Define a time window for analysis:
- For Spike Sorting Analysis: Set
session_typetospike_sorting, and create anEphysSessionProbeto store probe information, including the channel mapping. This triggers probe insertion detection automatically. For spike sorting, you will need to manually select the spike sorting algorithm and parameter set to run in the next step. - For LFP Analysis: Set
session_typetolfp, orboth(spike sorting and lfp analyses for the session selected). This automatically run the LFP analysis pipeline.
- For Spike Sorting Analysis: Set
- Create a
ClusteringTaskby selecting a spike-sorting algorithm and parameter set:- Go to works.datajoint.com →
Notebooktab - Run CREATE_new_clustering_paramset.ipynb to configure a new parameter set.
- Assign parameters to an
EphysSessionusing CREATE_new_clustering_task.ipynb. - The pipeline will automatically run the spike sorting process.
- Follow the download spike sorting results to retrieve results.
- Go to works.datajoint.com →
- Go to works.datajoint.com →
Notebooktab - Open EXPLORE_spike_sorting.ipynb to inspect processed ephys data.
- Go to works.datajoint.com →
Notebooktab - Open EXPLORE_LFP_analysis.ipynb to inspect processed LFP data.
- Request Axon credentials from the DataJoint support team.
- Set up your local machine (if you haven't already):
- Install the pipeline code.
- Configure axon settings (Cloud upload configuration).
- Download spike sorting results via the DOWNLOAD_spike_sorted_data.ipynb using either:
- Jupyter Notebook Server:
- Open a terminal or command prompt.
- Activate the
utah_organoidsenvironment withconda activate utah_organoids. - Ensure
Jupyteris installed in theutah_organoidsenvironment. If not, install it by runningconda install jupyter. - Navigate to the
utah_organoids/notebooksdirectory in the terminal. - Run
jupyter notebookin the terminal which will open the Jupyter notebook web interface. - Click on the notebook there (
DOWNLOAD_spike_sorted_data.ipynb) and follow the instructions to download results. - Note: to execute each code cell sequentially, press
Shift + Enteron your keyboard or click "Run". - Close the browser tab and stop Jupyter with
Ctrl + Cin the terminal when you are done with the upload and notebook.
- Visual Studio Code (VS Code):
- Install VS Code and the Python extension.
- Select the kernel for the notebook by clicking on the kernel name
utah_organoidsin the top right corner of the notebook. - Open the
DOWNLOAD_spike_sorted_data.ipynbnotebook in VS Code. - Click on the "Run Cell" button in the top right corner of each code cell to execute the code.
- Follow the instructions in the notebook to download spike sorting results.
- Jupyter Notebook Server:
For help, refer to the Documentation, Troubleshooting Guide, or contact the DataJoint support team.
If your work uses DataJoint Python, DataJoint Elements, or any integrated tools within the pipeline, please cite the respective manuscripts and Research Resource Identifiers (RRIDs).
Yatsenko D, Reimer J, Ecker AS, Walker EY, Sinz F, Berens P, Hoenselaar A, Cotton RJ, Siapas AS, Tolias AS.
DataJoint: managing big scientific data using MATLAB or Python. bioRxiv. 2015 Jan 1:031658.
DOI: 10.1101/031658
Resource Identification (RRID): SCR_014543
Yatsenko D, Walker EY, Tolias AS.
DataJoint: a simpler relational data model. arXiv:1807.11104. 2018 Jul 29.
DOI: 10.48550/arXiv.1807.11104
Resource Identification (RRID): SCR_014543
Yatsenko D, Nguyen T, Shen S, Gunalan K, Turner CA, Guzman R, Sasaki M, Sitonic D, Reimer J, Walker EY, Tolias AS.
DataJoint Elements: Data Workflows for Neurophysiology. bioRxiv. 2021 Jan 1.
DOI: 10.1101/2021.03.30.437358
Resource Identification (RRID): SCR_021894
- If your work uses SpikeInterface, please cite the respective manuscript.
- For other integrated tools within the pipeline, cite their respective manuscripts and RRIDs.