I want to map the" ts_normalized" surf_file to the AFNI volume domain(group level), however, the brick datasets in the generated “lh/rh_ts_normalized_sb_tlrc” file don’t have same number of volumes with the original “ts_normalized_tlrc” file.
foreach hemi (lh rh)
-f_p1_fr -0.2 -f_pn_fr 0.4
You seem to be missing an “end” for one of your loops.
What are the dimensions of your 1D file, that would have your data to map? THis will tell you about how many time points you have in your surface data set of choice. From the 3dSurf2Vol help file:
Note: With -sdata_1D, the first column of the file
should contain a node's index, and following columns are
that node's data. See the '-sdata_1D' option for more info.
Option -sdata takes NIML or GIFTI input which contain
node index information in their headers.
To find out the dimensions, for example, try:
1d_tool.py -infile "$hemi"_ts_normalized.1D -show_rows_cols
What are the dimensions of your << original “ts_normalized_tlrc” file. >>?
Check this with something like:
3dinfo -nv FILE
And finally, how did you generate the 1D surface files? What is your afni_proc.py command, for example? Did you select to remove TRs from teh beginning of the data set?