Changes for page SGA2 SP3 UC002 KR3.2 - Slow Wave Analysis Pipeline
Last modified by robing on 2022/03/25 09:55
Summary
-
Page properties (2 modified, 0 added, 0 removed)
-
Objects (1 modified, 2 added, 0 removed)
Details
- Page properties
-
- Author
-
... ... @@ -1,1 +1,1 @@ 1 -XWiki. denker1 +XWiki.robing - Content
-
... ... @@ -43,6 +43,8 @@ 43 43 44 44 This collab illustrates the usage of the [[Neo>>https://neo.readthedocs.io/en/stable/]] and [[Elephant>>https://elephant.readthedocs.io/en/latest/]] tools in performing data analysis with regards to the SGA2-SP3-UC002 KR3.2, investigating sleep, anesthesia, and the transition to wakefulness: see Chapter 1 and Figure 2 of SGA2[[ Deliverable D3.2.1.>>https://drive.ebrains.eu/smart-link/17ac0d6e-e050-4a49-8ca2-e223b70a3121/]], for an overview of the scientific motivations and a description of the UseCase workflow; Chapter 2 (same document) for an introduction to KR3.2; Chapter 3, for a description of the mice ECoG data sets; Chapter 5, about the Slow Wave Analysis Pipeline and Chapter 6 for the mice wide-field GECI data). For details on the datasets used in this collab, please see the References below. 45 45 46 +[[See the introduction video>>https://www.youtube.com/watch?v=uuAiY6HScM0]] 47 + 46 46 == How the pipeline works == 47 47 48 48 The design of the pipeline aims at interfacing a variety of general and specific analysis and processing steps in a flexible modular manner. Hence, it enables the pipeline to adapt to diverse types of data (e.g., electrical ECoG, or optical Calcium Imaging recordings) and to different analysis questions. This makes the analyses a) more reproducible and b) comparable amongst each other since they rely on the same stack of algorithms and any differences in the analysis are fully transparent. ... ... @@ -73,11 +73,13 @@ 73 73 Each stage has config files (//pipeline/<stage_name>/configs/config_<profile>.yaml//) to specify which analysis/processing blocks to execute and which parameters to use. General and specific information about the blocks and parameters can be found in the README and config files of each stage. There are preset configuration profiles for the benchmark datasets IDIBAPS ([[ECoG, anesthetized mouse>>https://kg.ebrains.eu/search/?facet_type[0]=Dataset&q=sanchez-vives#Dataset/2ead029b-bba5-4611-b957-bb6feb631396]]) and LENS ([[Calcium Imaging, anesthetized mouse>>https://kg.ebrains.eu/search/instances/Dataset/71285966-8381-48f7-bd4d-f7a66afa9d79]]). 74 74 75 75 * **Run the notebook** 76 -In the jupyter hub, navigate to //drive/My Libraries/My Library/ pipeline/showcase_notebooks/run_snakemake_in_collab.ipynb//, or where you copied the//pipeline//folderto.77 -Follow the notebook to install the required packages into your Python kernel, set the output path, and execute the pipeline with snakemake 78 +In the jupyter hub, navigate to //drive/My Libraries/My Library/run_snakemake_in_collab.ipynb//, or where you copied the file to. 79 +Follow the notebook to install the required packages into your Python kernel, set the output path, and execute the pipeline with snakemake. 78 78 79 79 === ii) Local execution === 80 80 83 +//tested only with Mac OS and Linux!// 84 + 81 81 * **Get the code** 82 82 The source code of the pipeline is available via Github: [[INM-6/wavescalephant>>https://github.com/INM-6/wavescalephant]] and can be cloned to your machine ([[how to get started with Github>>https://guides.github.com/activities/hello-world/]]). 83 83 ... ... @@ -86,6 +86,10 @@ 86 86 In the wavescalephant git repository, there is an environment file ([[pipeline/environment.yaml>>https://drive.ebrains.eu/smart-link/1a0b15bb-be87-46ee-b838-4734bc320d20/]]) specifying the required packages and versions. To build the environment, we recommend using conda ([[how to get started with conda>>https://docs.conda.io/projects/conda/en/latest/user-guide/getting-started.html]]). 87 87 ##conda env create ~-~-file environment.yaml 88 88 conda activate wavescalephant_env## 93 + 94 +Make sure that neo and elephant were installed as their Github development version, and if necessary add them manually to the environment. 95 +##pip install git+https:~/~/github.com/NeuralEnsemble/elephant.git 96 +pip install git+https:~/~/github.com/NeuralEnsemble/python-neo.git## 89 89 90 90 ))) 91 91 * **Edit the settings** ... ... @@ -113,6 +113,18 @@ 113 113 114 114 All results are stored in the path specified in the //settings.py// file. The folder structure reflects the structuring of the pipeline into stages and blocks. All intermediate results are stored as //.nix// files using the [[Neo data format>>https://neo.readthedocs.io/en/stable/]] and can be loaded with ##neo.NixIO('/path/to/file.nix').read_block()##. Additionally, most blocks produce a figure, and each stage a report file, to give an overview of the execution log, parameters, intermediate results, and to help with debugging. The final stage (//stage05_wave_characterization//) stores the results as[[ //pandas.DataFrames//>>https://pandas.pydata.org/]] in //.csv// files, separately for each measure as well as in a combined dataframe for all measures. 115 115 124 +**Examples of the output figures (for IDIBAPS dataset)** 125 + 126 +* Stage 01 - [[example signal traces and metadata>>https://drive.ebrains.eu/smart-link/cf2fa914-260d-4d61-a2da-03ea07b7f9be/]] 127 +* Stage 02 - [[background substraction>>https://drive.ebrains.eu/smart-link/586d2f3c-591b-4dfb-94ee-8c0e28050dc4/]] 128 +* Stage 02 - [[logMUA estimation>>https://drive.ebrains.eu/smart-link/c92e4b0c-0938-44e8-9f8d-00522796b2fd/]] 129 +* Stage 02 - [[processed signal trace>>https://drive.ebrains.eu/smart-link/26ed27c6-de56-4b48-a57b-f70aab629197/]] 130 +* Stage 03 - [[amplitude distribution>>https://drive.ebrains.eu/smart-link/8ba80293-ba75-4a37-8a8f-05d44cf6f65c/]] 131 +* Stage 03 - [[UP state detection>>https://drive.ebrains.eu/smart-link/ab172be0-178e-4153-a3e6-b4bace32dd50/]] 132 +* Stage 04 - [[trigger clustering>>https://drive.ebrains.eu/smart-link/4a1f0169-8b43-49ce-80c8-f2fa0f4d50d3/]] 133 +* Stage 05 - [[planar velocities>>https://drive.ebrains.eu/smart-link/f4de8073-cb40-47a7-bc82-f97d36dbae25/]] 134 +* Stage 05 - [[directionality>>https://drive.ebrains.eu/smart-link/5485032d-0121-4cde-9ea2-3e0af3f12178/]] 135 + 116 116 == Outlook == 117 117 118 118 * Using the **KnowledgeGraph API **to insert data directly from the Knowledge Graph into the pipeline and also register and store the corresponding results as Analysis Objects. Such Analysis Objects are to incorporate **Provenance Tracking, **using [[fairgraph>>https://github.com/HumanBrainProject/fairgraph]],** **to record the details of the processing and analysis steps.
- Collaboratory.Apps.Collab.Code.CollabClass[0]
-
- Public
-
... ... @@ -1,1 +1,1 @@ 1 - No1 +Yes
- XWiki.XWikiRights[3]
-
- Allow/Deny
-
... ... @@ -1,0 +1,1 @@ 1 +Allow - Levels
-
... ... @@ -1,0 +1,1 @@ 1 +view - Users
-
... ... @@ -1,0 +1,1 @@ 1 +XWiki.XWikiGuest
- XWiki.XWikiRights[4]
-
- Allow/Deny
-
... ... @@ -1,0 +1,1 @@ 1 +Allow - Groups
-
... ... @@ -1,0 +1,1 @@ 1 +XWiki.XWikiAllGroup - Levels
-
... ... @@ -1,0 +1,1 @@ 1 +view