{"id":13676355,"url":"https://github.com/bids-apps/CPAC","last_synced_at":"2025-04-29T06:31:50.968Z","repository":{"id":48427721,"uuid":"64773036","full_name":"bids-apps/CPAC","owner":"bids-apps","description":"BIDS Application for the Configurable Pipeline for the Analysis of Connectomes (C-PAC)","archived":false,"fork":false,"pushed_at":"2024-11-04T08:30:34.000Z","size":68580,"stargazers_count":14,"open_issues_count":11,"forks_count":18,"subscribers_count":6,"default_branch":"master","last_synced_at":"2024-11-11T18:41:08.373Z","etag":null,"topics":["bids","bidsapp","mri","preprocessing"],"latest_commit_sha":null,"homepage":"","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"apache-2.0","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/bids-apps.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2016-08-02T16:20:58.000Z","updated_at":"2024-11-04T08:30:39.000Z","dependencies_parsed_at":"2024-04-08T22:31:59.859Z","dependency_job_id":"20bd1cb3-4b67-4231-b771-12718e0331e9","html_url":"https://github.com/bids-apps/CPAC","commit_stats":null,"previous_names":[],"tags_count":44,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/bids-apps%2FCPAC","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/bids-apps%2FCPAC/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/bids-apps%2FCPAC/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/bids-apps%2FCPAC/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/bids-apps","download_url":"https://codeload.github.com/bids-apps/CPAC/tar.gz/refs/heads/master","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":251450656,"owners_count":21591407,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["bids","bidsapp","mri","preprocessing"],"created_at":"2024-08-02T13:00:23.754Z","updated_at":"2025-04-29T06:31:45.951Z","avatar_url":"https://github.com/bids-apps.png","language":"Python","funding_links":[],"categories":["BIDS Apps"],"sub_categories":["others"],"readme":"# C-PAC BIDS Application\n\n## Documentation\nExtensive information can be found in the [C-PAC User Guide](http://fcp-indi.github.io/docs/user/index.html).\n\n## Description\nThe Configurable Pipeline for the Analysis of Connectomes [C-PAC](http://fcp-indi.github.io) is a software for\nperforming high-throughput preprocessing and analysis of functional connectomes data using high-performance computers.\nC-PAC is implemented in Python using the Nipype pipelining [[1](#nipype)] library to efficiently combine tools from AFNI\n[[2](#afni)], ANTS [[3](#ants)], and FSL [[4](#fsl)] to achieve high quality and robust automated processing. This\ndocker container, when built, is an application for performing participant level analyses. Future releases will include\ngroup-level analyses, when there is a BIDS standard for handling derivatives and group models.\n\n### A note about versioning\nC-PAC BIDS Apps version tags are composed of the C-PAC version followed by an underscore and then the version of the\ncontainer. The container version restarts for every new C-PAC version and is a single integer that reflects the\nmodification number of the build. For example ```v1.0.1a_5``` corresponds to the 5th build of the container for C-PAC\nversion v1.0.1a.\n\n### Usage notes\n\n1. You can either perform a custom processing using a YAML configuration file or use the default processing pipeline. A\nGUI can be invoked to assist in pipeline custimization by specifying `GUI` command line arguement (this currently only\nworks for Singularity containers).\n\n2. The default behavior is to read in data that is organized in the BIDS format. This includes data that is in Amazon\nAWS S3 by using the format `s3://\u003cbucket_name\u003e/\u003cbids_dir\u003e` for the `bids_dir` command line argument. Outputs can be\nwritten to S3 using the same format for the output_dir. Credentials for accessing these buckets can be specified on the\ncommand line (using `--aws_input_creds` or `--aws_output_creds`).\n\n3. Non-BIDS organized data can processed using a C-PAC data configuration yaml file. This file can be generated using\nthe C-PAC GUI (start the app with the `GUI` argument, also see instructions [below](#GUI)) or can be created using other means, please refer to [CPAC\ndocumentation](http://fcp-indi.github.io/docs/user/subject_list_config.html) for more information.\n\n4. When the app is run, a data configuration file is written to the working directory. This file can be passed into\nsubsequent runs, which avoids the overhead of re-parsing the BIDS input directory on each run (i.e. for cluster or\ncloud runs). These files can be generated without executing the C-PAC pipeline using the `test_run` command line\nargument.\n\n5. The `participant_label` and `participant_ndx` arguments allow the user to specify which of the many datasets should\nbe processed, this are useful when parallelizing the run of multiple participants.\n\n### Default configuration\nThe default processing pipeline performs fMRI processing using four strategies, with and without global signal\nregression, with and without bandpass filtering.\n\nAnatomical processing begins with conforming the data to RPI orientation and removing orientation header information\nthat will interfere with further processing. A  non-linear transform between skull-on images and a 2mm MNI brain-only\ntemplate  are calculated using ANTs [[3](#ants)]. Images are them skull-stripped using AFNI's `3dSkullStrip` [[5](#bet)]\nand subsequently segmented into WM, GM, and CSF using FSL’s `fast` tool [[6](#fast)]. The resulting WM mask was\nmultiplied by a WM prior map that was transformed into individual space using the inverse of the linear transforms\npreviously calculated during the ANTs procedure. A CSF mask was multiplied by a ventricle map derived from the\nHarvard-Oxford atlas distributed with FSL [[4](#fsl)]. Skull-stripped images and grey matter tissue maps are written\ninto MNI space at 2mm resolution.\n\nFunctional preprocessing begins with resampling the data to RPI orientation, and slice timing correction. Next, motion\ncorrection is performed using a two-stage approach in which the images are first coregistered to the mean fMRI and then\na new mean is calculated and used as the target for a second coregistration (AFNI `3dvolreg` [[2](#afni)]). A 7 degree of\nfreedom linear transform between the mean fMRI and the structural image is calculated using FSL’s implementation of\nboundary-based registration [[7](#bbr)]. Nuisance variable regression (NVR) is performed on motion corrected data using\na 2nd order polynomial, a 24-regressor model of motion [[8](#friston24)], 5 nuisance signals, identified via principal\ncomponents analysis of signals obtained from white matter (CompCor, [[9](#compcor)]), and mean CSF signal. WM and CSF\nsignals were extracted using the previously described masks after transforming the fMRI data to match them in 2mm space\nusing the inverse of the linear fMRI-sMRI transform. The NVR procedure is performed twice, with and without the\ninclusion of the global signal as a nuisance regressor. The residuals of the NVR procedure are processed with and\nwithout bandpass filtering (0.001Hz \u003c f \u003c 0.1Hz), written into MNI space at 3mm resolution and subsequently smoothed\nusing a 6mm FWHM kernel.\n\nSeveral different individual level analysis are performed on the fMRI data including:\n\n- **Amplitude of low frequency fluctuations (alff) [[10](#alff)]**: the variance of each voxel is calculated after\nbandpass filtering in original space and subsequently written into MNI space at 2mm resolution and spatially smoothed\nusing a 6mm FWHM kernel.\n- **Fractional amplitude of low frequency fluctuations (falff) [[11](#falff)]**: Similar to alff except that the\nvariance of the bandpassed signal is divided by the total variance (variance of non-bandpassed signal.\n- **Regional homogeniety (ReHo) [[12](#reho)]**: a simultaneous Kendalls correlation is calculated between each\nvoxel's time course and the time courses of the 27 voxels that are face, edge, and corner touching the voxel. ReHo is\ncalculated in original space and subsequently written into MNI space at 2mm resolution and spatially smoothed using\na 6mm FWHM kernel.\n- **Voxel mirrored homotopic connectivity (VMHC) [[13](#vmhc)]**: an non-linear transform is calculated between the\nskull-on anatomical data and a symmetric brain template in 2mm space. Using this transform, processed fMRI data are\nwritten in to symmetric MNI space at 2mm and the correlation between each voxel and its analog in the contralateral\nhemisphere is calculated. The Fisher transform is applied to the resulting values, which are then spatially smoothed\nusing a 6mm FWHM kernel.\n- **Weighted and binarized degree centrality (DC) [[14](#degree)]**: fMRI data is written into MNI space at 2mm\nresolution and spatially smoothed using a 6mm FWHM kernel. The voxel x voxel similarity matrix is calculated by the\ncorrelation between every pair of voxel time courses and then thresholded so that only the top 5% of correlations\nremain. For each voxel, binarized DC is the number of connections that remain for the voxel after thresholding and\nweighted DC is the average correlation coefficient across the remaining connections.\n- **Eigenvector centrality (EC) [[15](#eigen)]**: fMRI data is written into MNI space at 2mm resolution and spatially\nsmoothed using a 6mm FWHM kernel. The voxel x voxel similarity matrix is calculated by the correlation between every\npair of voxel time courses and then thresholded so that only the top 5% of correlations remain. Weighted EC is\ncalculated from the eigenvector corresponding to the largest eigenvalue from an eigenvector decomposition of the\nresulting similarity. Binarized EC, is the first eigenvector of the similarity matrix after setting the non-zero values\nin the resulting matrix are set to 1.\n- **Local functional connectivity density (lFCD) [[16](#lfcd)]**: fMRI data is written into MNI space at 2mm resolution\nand spatially smoothed using a 6mm FWHM kernel. For each voxel, lFCD corresponds to the number of contiguous voxels that\nare correlated with the voxel above 0.6 (r\u003e0.6). This is similar to degree centrality, except only voxels that it only\nincludes the voxels that are directly connected to the seed voxel.\n- **10 intrinsic connectivity networks (ICNs) from dual regression [[17](#dr)]**: a template including 10 ICNs from a\nmeta-analysis of resting state and task fMRI data [[18](#icn10)] is spatially regressed against the processed fMRI data\nin MNI space. The resulting time courses are entered into a multiple regression with the voxel data in original space to\ncalculate individual representations of the 10 ICNs. The resulting networks are written into MNI space at 2mm and then\nspatially smoothed using a 6mm FWHM kernel.\n- **Seed correlation analysis (SCA)**: preprocessed fMRI data is to match template that includes 160 regions of interest\ndefined from a meta-analysis of different task results [[19](#do160)]. A time series is calculated for each region from\nthe mean of all intra-ROI voxel time series. A seperate functional connectivity map is calculated per ROI by correlating\nits time course with the time courses of every other voxel in the brain. Resulting values are Fisher transformed,\nwritten into MNI space at 2mm resolution, and then spatiall smoothed using a 6mm FWHM kernel.\n- **Time series extraction**: similar the procedure used for time series analysis, the preprocessed functional data is\nwritten into MNI space at 2mm and then time series for the various atlases are extracted by averaging within region\nvoxel time courses. This procedure was used to generate summary time series for the automated anatomic labelling atlas\n[[20](#aal)], Eickhoff-Zilles atlas [[21](#ez)], Harvard-Oxford atlas [[22](#ho)], Talaraich and Tournoux atlas\n[[23](#tt)], 200 and 400 regions from the spatially constrained clustering voxel timeseries [[24](#cc)], and 160 ROIs\nfrom a meta-analysis of task results [[19](#do160)]. Time series for 10 ICNs were extracted using spatial regression.\n\n## Usage\nThis App has the following command line arguments:\n\n    usage: run.py [-h] [--pipeline_file PIPELINE_FILE]\n                  [--data_config_file DATA_CONFIG_FILE]\n                  [--aws_input_creds AWS_INPUT_CREDS]\n                  [--aws_output_creds AWS_OUTPUT_CREDS] [--n_cpus N_CPUS]\n                  [--mem_mb MEM_MB] [--mem_gb MEM_GB] [--save_working_dir]\n                  [--participant_label PARTICIPANT_LABEL [PARTICIPANT_LABEL ...]]\n                  [--participant_ndx PARTICIPANT_NDX]\n                  bids_dir output_dir {participant,group,test_config,GUI}\n\n    C-PAC Pipeline Runner\n\n    positional arguments:\n      bids_dir              The directory with the input dataset formatted\n                            according to the BIDS standard. Use the format\n                            s3://bucket/path/to/bidsdir to read data directly from\n                            an S3 bucket. This may require AWS S3 credentials\n                            specificied via the --aws_input_creds option.\n      output_dir            The directory where the output files should be stored.\n                            If you are running group level analysis this folder\n                            should be prepopulated with the results of the\n                            participant level analysis. Us the format\n                            s3://bucket/path/to/bidsdir to write data directly to\n                            an S3 bucket. This may require AWS S3 credentials\n                            specificied via the --aws_output_creds option.\n      {participant,group,test_config,GUI}\n                            Level of the analysis that will be performed. Multiple\n                            participant level analyses can be run independently\n                            (in parallel) using the same output_dir. GUI will open\n                            the CPAC gui (currently only works with singularity)\n                            and test_config will run through the entire\n                            configuration process but will not execute the\n                            pipeline.\n\n    optional arguments:\n      -h, --help            show this help message and exit\n      --pipeline_file PIPELINE_FILE\n                            Name for the pipeline configuration file to use\n      --data_config_file DATA_CONFIG_FILE\n                            Yaml file containing the location of the data that is\n                            to be processed. Can be generated from the CPAC gui.\n                            This file is not necessary if the data in bids_dir is\n                            organized according to the BIDS format. This enables\n                            support for legacy data organization and cloud based\n                            storage. A bids_dir must still be specified when using\n                            this option, but its value will be ignored.\n      --aws_input_creds AWS_INPUT_CREDS\n                            Credentials for reading from S3. If not provided and\n                            s3 paths are specified in the data config we will try\n                            to access the bucket anonymously\n      --aws_output_creds AWS_OUTPUT_CREDS\n                            Credentials for writing to S3. If not provided and s3\n                            paths are specified in the output directory we will\n                            try to access the bucket anonymously\n      --n_cpus N_CPUS       Number of execution resources available for the\n                            pipeline\n      --mem_mb MEM_MB       Amount of RAM available to the pipeline in megabytes.\n                            Included for compatibility with BIDS-Apps standard,\n                            but mem_gb is preferred\n      --mem_gb MEM_GB       Amount of RAM available to the pipeline in gigabytes.\n                            if this is specified along with mem_mb, this flag will\n                            take precedence.\n      --save_working_dir    Save the contents of the working directory.\n      --participant_label PARTICIPANT_LABEL [PARTICIPANT_LABEL ...]\n                            The label of the participant that should be analyzed.\n                            The label corresponds to sub-\u003cparticipant_label\u003e from\n                            the BIDS spec (so it does not include \"sub-\"). If this\n                            parameter is not provided all subjects should be\n                            analyzed. Multiple participants can be specified with\n                            a space separated list. To work correctly this should\n                            come at the end of the command line\n      --participant_ndx PARTICIPANT_NDX\n                            The index of the participant that should be analyzed.\n                            This corresponds to the index of the participant in\n                            the subject list file. This was added to make it\n                            easier to accomodate SGE array jobs. Only a single\n                            participant will be analyzed. Can be used with\n                            participant label, in which case it is the index into\n                            the list that follows the particpant_label flag.\n\n### To run it in participant level mode (for one participant):\n\n    docker run -i --rm \\\n        -v /tmp:/scratch \\\n        -v /Users/filo/data/ds005:/bids_dataset \\\n        -v /Users/filo/outputs:/outputs \\\n\t\tbids/cpac \\\n\t\t/bids_dataset /outputs participant --participant_label 01\n\n### Running the \u003ca id=\"GUI\"\u003eGUI\u003c/a\u003e (requires mapping your X socket)\n\n#### Running docker container on Linux\n\n1. Start the docker container, mapping the X socket (change /Users/filo to a local directory on your computer)\n\n```bash\n    docker run -i --rm \\\n        --privileged \\\n        -e DISPLAY=$DISPLAY \\\n        -v /tmp/.X11-unix:/tmp/.X11-unix \\\n        -v /tmp:/scratch \\\n        -v /Users/filo/data/ds005:/bids_dataset \\\n        -v /Users/filo/outputs:/outputs \\\n        bids/cpac \\\n        /bids_dataset /outputs GUI\n```\n\n#### Running docker container on Mac OSX\n\n1. [Install XQuartz](https://www.xquartz.org/)\n\n2. Start XQuartz (from terminal)\n\n```bash\n    open -a XQuartz\n```\n\n3. Enable XQuartz connections from network clients\n\nXQuartz -\u003e preferences -\u003e security -\u003e \"Allow connections from network clients\"\n\n4. Get your ip address (e.g., might have to change eth0 to match the name of your network interface.)\n\n```bash\n    ip=$(ifconfig en0 | grep inet | awk '$1==\"inet\" {print $2}')\n\n```\n\n5. Tell xhost to accept connections from the localhost\n\n```bash\nxhost + ${ip}\n```\n\n6. Start the docker container, mapping the X socket (change /Users/filo to a local directory on your computer)\n\n```bash\n    docker run -i --rm \\\n        --privileged \\\n        -e DISPLAY=$ip:0 \\\n        -v /tmp/.X11-unix:/tmp/.X11-unix \\\n        -v /tmp:/scratch \\\n        -v /Users/filo/data/ds005:/bids_dataset \\\n        -v /Users/filo/outputs:/outputs \\\n\t\tbids/cpac \\\n\t\t/bids_dataset /outputs GUI\n```\n\n#### Running singularity container on Linux\n\n1. Start the docker container (it just works!, provided you change /Users/filo to a local directory on your computer)\n\n```bash\n    singularity run \\\n        -B /home/ubuntu:/mnt \\\n        -B /mnt:/scratch \\\n        -B /Users/filo/data/ds005:/bids_dataset \\\n        -B /Users/filo/outputs:/outputs \\\n        /home/ubuntu/workspace/container_build/singularity_images/cpac_latest.img \\\n        /bids_dataset \\\n        /outputs\\\n        GUI\n```\n\n### To convert the Docker container to a [Singularity](http://singularity.lbl.gov/) container :\n\n    docker run --privileged -ti --rm  \\\n        -v /var/run/docker.sock:/var/run/docker.sock \\\n        -v /home/srycajal/singularity_images:/output \\\n        filo/docker2singularity \\\n        bids/cpac\n\n### Example submit script for running as a Singularity container on sun grid engine:\n\n    #! /bin/bash\n    ## SGE batch file - bgsp\n    #$ -S /bin/bash\n    ## bgsp is the jobname and can be changed\n    #$ -N bgsp\n    ## execute the job using the mpi_smp parallel enviroment and 8 cores per job\n    #$ -pe mpi_smp 8\n    ## create an array of 1112 jobs\n    #$ -t 1-1112\n    #$ -V\n    ## change the following working directory to a persistent directory that is\n    ## available on all nodes, this is were messages printed by the app (stdout\n    ## and stderr) will be stored\n    #$ -wd /home/ubuntu/workspace/cluster_files\n\n    sudo chmod 777 /mnt\n    mkdir -p /mnt/log/reports\n\n    sge_ndx=$(( SGE_TASK_ID - 1 ))\n\n    # random sleep so that jobs dont start at _exactly_ the same time\n    sleep $(( $SGE_TASK_ID % 10 ))\n\n    singularity run -B /home/ubuntu:/mnt -B /mnt:/scratch \\\n      /home/ubuntu/workspace/container_build/singularity_images/cpac_latest.img \\\n      --n_cpus 8 --mem 12 \\\n      --aws_input_creds /mnt/workspace/cluster_files/s3-keys.csv \\\n      --aws_output_creds /mnt/workspace/cluster_files/s3-keys.csv \\\n      --data_config_file /mnt/workspace/cluster_files/bgsp_data_config.yml \\\n      s3://fcp-indi/data/Projects/BrainGenomicsSuperstructProject/orig_bids/ \\\n      s3://fcp-indi/data/Projects/BrainGenomicsSuperstructProject/cpac_out/ \\\n      participant --participant_ndx ${sge_ndx}\n\n#### Notes:\n1. With the exception of your home directory, which is mounted from the local filesystem, the filesystem in Singularity\ncontainers is read-only. Files can be easily transferred in and out of the container by mapping local directories to\ndirectories inside the container using the `-B from:to` command line argument, where the `from` dir is mapped to `to`.\nWhen using mapped directories, remember that the paths specified on the command line are in relation to the directory\ninside the container (e.g. the `to` directory).\n\n2. Unless the `--save_working_dir` flag is set, the C-PAC app will use the `/scratch` directory for intermediary files.\nSince this directory is write protected, a directory from the local filesystem must be mapped to `/scratch` for the\npipeline to run successfully. This directory should be large  enough to hold all of the intermediary files for the\ndatasets that are processed in parallel, as a rule of thumb we suggest 3 GB per dataset. Unless the `--save_working_dir`\nflag is set, the working directory will be deleted when the pipeline has completed.\n\n3. Use the `--save_working_dir` flag to retain all intermediary files, which can be useful for debugging. In this case,\nthe intermediary files will be saved in the `working_dir` subdirectory of the user specified `output` directory. This\nwill require about 3GB per dataset, but may require more for multiple or very long fMRI scans.\n\n## Reporting errors and getting help\nPlease report errors on the [C-PAC github page issue tracker](https://github.com/FCP-INDI/C-PAC/issues). Please use the\n[C-PAC google group](https://groups.google.com/forum/#!forum/cpax_forum) for help using C-PAC and this application.\n\n\n## Acknowledgements\nWe currently have a publication in preparation, in the meantime please cite our poster from INCF:\n\n    Craddock C, Sikka S, Cheung B, Khanuja R, Ghosh SS, Yan C, Li Q, Lurie D, Vogelstein J, Burns R, Colcombe S,\n    Mennes M, Kelly C, Di Martino A, Castellanos FX and Milham M (2013). Towards Automated Analysis of Connectomes:\n    The Configurable Pipeline for the Analysis of Connectomes (C-PAC). Front. Neuroinform. Conference Abstract:\n    Neuroinformatics 2013. doi:10.3389/conf.fninf.2013.09.00042\n\n    @ARTICLE{cpac2013,\n        AUTHOR={Craddock, Cameron  and  Sikka, Sharad  and  Cheung, Brian  and  Khanuja, Ranjeet  and  Ghosh, Satrajit S\n            and Yan, Chaogan  and  Li, Qingyang  and  Lurie, Daniel  and  Vogelstein, Joshua  and  Burns, Randal  and\n            Colcombe, Stanley  and  Mennes, Maarten  and  Kelly, Clare  and  Di Martino, Adriana  and  Castellanos,\n            Francisco Xavier  and  Milham, Michael},\n        TITLE={Towards Automated Analysis of Connectomes: The Configurable Pipeline for the Analysis of Connectomes (C-PAC)},\n        JOURNAL={Frontiers in Neuroinformatics},\n        YEAR={2013},\n        NUMBER={42},\n        URL={http://www.frontiersin.org/neuroinformatics/10.3389/conf.fninf.2013.09.00042/full},\n        DOI={10.3389/conf.fninf.2013.09.00042},\n        ISSN={1662-5196}\n    }\n\n## References\n\n\u003ca id=\"nipype\"\u003e1.\u003c/a\u003e Gorgolewski, K., Burns, C.D., Madison, C., Clark, D., Halchenko, Y.O., Waskom, M.L., Ghosh, S.S.: Nipype: A flexible, lightweight and extensible neuroimaging data processing framework in python. Front. Neuroinform. 5\n(2011). doi:10.3389/fninf.2011.00013\n\n\u003ca id=\"afni\"\u003e2.\u003c/a\u003e Cox, R.W., Jesmanowicz, A.: Real-time 3d image registration for functional mri. Magn Reson Med 42(6),\n1014–8 (1999)\n\n\u003ca id=\"ants\"\u003e3.\u003c/a\u003e Avants, B., Epstein, C., Grossman, M., Gee, J.: Symmetric diffeomorphic image registration with\ncross-correlation: Evaluating automated labeling of elderly and neurodegenerative brain. Medical Image Analysis\n12(1), 26–41 (2008). doi:10.1016/j.media.2007.06.004\n\n\u003ca id=\"fsl\"\u003e4.\u003c/a\u003e Smith, S.M., Jenkinson, M., Woolrich, M.W., Beckmann, C.F., Behrens, T.E.J., Johansen-Berg, H., Bannister,\nP.R., Luca, M.D., Drobnjak, I., Flitney, D.E., Niazy, R.K., Saunders, J., Vickers, J., Zhang, Y., Stefano, N.D.,\nBrady, J.M., Matthews, P.M.: Advances in functional and structural mr image analysis and implementation as\nfsl. NeuroImage 23, 208–219 (2004). doi:10.1016/j.neuroimage.2004.07.051\n\n\u003ca id=\"bet\"\u003e5.\u003c/a\u003e  Smith, S.M.: Fast robust automated brain extraction. Human Brain Mapping 17(3), 143–155 (2002). doi:10.1002/hbm.10062\n\n\u003ca id=\"fast\"\u003e6.\u003c/a\u003e Zhang, Y., Brady, M., Smith, S.: Segmentation of brain mr images through a hidden markov random field\nmodel and the expectation-maximization algorithm. IEEE Transactions on Medical Imaging 20(1), 45–57\n(2001). doi:10.1109/42.906424\n\n\u003ca id=\"bbr\"\u003e7.\u003c/a\u003e Greve, D.N., Fischl, B.: Accurate and robust brain image alignment using boundary-based registration.\nNeuroImage 48(1), 63–72 (2009). doi:10.1016/j.neuroimage.2009.06.060\n\n\u003ca id=\"friston24\"\u003e8.\u003c/a\u003e Friston, K.J., Williams, S., Howard, R., Frackowiak, R.S., Turner, R.: Movement-related effects in fmri time-series. Magn Reson Med 35(3), 346–55 (1996)\n\n\u003ca id=\"compcor\"\u003e9.\u003c/a\u003e Behzadi, Y., Restom, K., Liau, J., Liu, T.T.: A component based noise correction method (compcor) for bold and perfusion based fmri. NeuroImage 37(1), 90–101 (2007). doi:10.1016/j.neuroimage.2007.04.042\n\n\u003ca id=\"alff\"\u003e10.\u003c/a\u003e Zang, Y.-F., He, Y., Zhu, C.-Z., Cao, Q.-J., Sui, M.-Q., Liang, M., Tian, L.-X., et al. (2007). Altered baseline brain activity in children with ADHD revealed by resting-state functional MRI. Brain \u0026 development, 29(2), 83–91.\n\n\u003ca id=\"falff\"\u003e11.\u003c/a\u003e Zou, Q.-H., Zhu, C.-Z., Yang, Y., Zuo, X.-N., Long, X.-Y., Cao, Q.-J., Wang, Y.-F., et al. (2008). An improved approach to detection of amplitude of low-frequency fluctuation (ALFF) for resting-state fMRI: Fractional ALFF. Journal of neuroscience methods, 172(1), 137–141.\n\n\u003ca id=\"reho\"\u003e12.\u003c/a\u003e Zang, Y., Jiang, T., Lu, Y., He, Y., Tian, L., 2004. Regional homogeneity approach to fMRI data analysis. Neuroimage 22, 394-400.\n\n\u003ca id=\"vmhc\"\u003e13.\u003c/a\u003e Stark, D. E., Margulies, D. S., Shehzad, Z. E., Reiss, P., Kelly, A. M. C., Uddin, L. Q., Gee, D. G., et al. (2008). Regional variation in interhemispheric coordination of intrinsic hemodynamic fluctuations. The Journal of Neuroscience, 28(51), 13754–13764.\n\n\u003ca id=\"degree\"\u003e14.\u003c/a\u003e Buckner RL, Sepulcre J, Talukdar T, Krienen FM, Liu H, Hedden T, Andrews-Hanna JR, Sperling RA, Johnson KA. 2009. Cortical hubs revealed by intrinsic functional connectivity: mapping, assessment of stability, and relation to Alzheimer’s disease. J Neurosci. 29:1860–1873.\n\n\u003ca id=\"eigen\"\u003e15.\u003c/a\u003e Lohmann G, Margulies DS, Horstmann A, Pleger B, Lepsien J, Goldhahn D, Schloegl H, Stumvoll M, Villringer A, Turner R. 2010. Eigenvector centrality mapping for analyzing connectivity patterns in fMRI data of the human brain. PLoS One. 5:e10232\n\n\u003ca id=\"lfcd\"\u003e16.\u003c/a\u003e Tomasi D, Volkow ND. 2010. Functional connectivity density mapping. PNAS. 107(21):9885-9890.\n\n\u003ca id=\"dr\"\u003e17.\u003c/a\u003e C.F. Beckmann, C.E. Mackay, N. Filippini, and S.M. Smith. Group comparison of resting-state FMRI data using multi-subject ICA and dual regression. OHBM, 2009.\n\n\u003ca id=\"icn10\"\u003e18.\u003c/a\u003e Smith, S. M., Fox, P. T., Miller, K. L., Glahn, D. C., Fox, P. M., Mackay, C. E., et al. (2009). Correspondence of the brain’s functional architecture during activation and rest. Proceedings of the National Academy of Sciences of the United States of America, 106(31), 13040–13045. doi:10.1073/pnas.0905267106\n\n\u003ca id=\"do160\"\u003e19.\u003c/a\u003e Dosenbach, N. U. F., Nardos, B., Cohen, A. L., Fair, D. a, Power, J. D., Church, J. a, … Schlaggar, B. L. (2010). Prediction of individual brain maturity using fMRI. Science (New York, N.Y.), 329(5997), 1358–61. http://doi.org/10.1126/science.1194144\n\n\u003ca id=\"aal\"\u003e20.\u003c/a\u003e Tzourio-Mazoyer, N., Landeau, B., Papathanassiou, D., Crivello, F., Etard, O., Delcroix, N., … Joliot, M. (2002). Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. NeuroImage, 15(1), 273–89. http://doi.org/10.1006/nimg.2001.0978\n\n\u003ca id=\"ez\"\u003e21.\u003c/a\u003e Eickhoff, S. B., Stephan, K. E., Mohlberg, H., Grefkes, C., Fink, G. R., Amunts, K., \u0026 Zilles, K. (2005). A new SPM toolbox for combining probabilistic cytoarchitectonic maps and functional imaging data. NeuroImage, 25(4), 1325–35. http://doi.org/10.1016/j.neuroimage.2004.12.034\n\n\u003ca id=\"ho\"\u003e22.\u003c/a\u003e Harvard-Oxford cortical and subcortical structural atlases, http://fsl.fmrib.ox.ac.uk/fsl/fslwiki/Atlases\n\n\u003ca id=\"tt\"\u003e23.\u003c/a\u003e Lancaster, J. L., Woldorff, M. G., Parsons, L. M., Liotti, M., Freitas, C. S., Rainey, L., … Fox, P. T. (2000). Automated Talairach atlas labels for functional brain mapping. Human Brain Mapping, 10(3), 120–31. Retrieved from http://www.ncbi.nlm.nih.gov/pubmed/10912591\n\n\u003ca id=\"cc\"\u003e24.\u003c/a\u003e Craddock, R. C., James, G. A., Holtzheimer, P. E., Hu, X. P., \u0026 Mayberg, H. S. (2011). A whole brain fMRI atlas generated via spatially constrained spectral clustering. Human Brain Mapping, 0(July 2010). http://doi.org/10.1002/hbm.21333\n","project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fbids-apps%2FCPAC","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fbids-apps%2FCPAC","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fbids-apps%2FCPAC/lists"}