{"id":54,"date":"2019-06-05T20:18:24","date_gmt":"2019-06-05T20:18:24","guid":{"rendered":"http:\/\/flux.cs.queensu.ca\/medi\/?page_id=54"},"modified":"2024-07-27T23:53:16","modified_gmt":"2024-07-27T23:53:16","slug":"publications","status":"publish","type":"page","link":"https:\/\/labs.cs.queensu.ca\/medilab\/publications\/","title":{"rendered":"Publications"},"content":{"rendered":"<div class=\"teachpress_pub_list\"><form name=\"tppublistform\" method=\"get\"><a name=\"tppubs\" id=\"tppubs\"><\/a><div class=\"tp_search_input\"><input name=\"tsr\" id=\"tp_search_input_field\" type=\"search\" placeholder=\"Enter search word\" value=\"\" tabindex=\"1\"\/><\/div><div class=\"teachpress_filter\"><select class=\"block\" title=\"All years\" name=\"yr\" id=\"yr\" tabindex=\"2\">\r\n                   <option value=\"\">All years<\/option>\r\n                   <option value=\"2025\" >2025<\/option><option value=\"2024\" >2024<\/option><option value=\"2023\" >2023<\/option><option value=\"2022\" >2022<\/option><option value=\"2021\" >2021<\/option><option value=\"2020\" >2020<\/option><option value=\"2019\" >2019<\/option><option value=\"2018\" >2018<\/option><option value=\"2017\" >2017<\/option><option value=\"2016\" >2016<\/option><option value=\"2015\" >2015<\/option><option value=\"2014\" >2014<\/option><option value=\"2013\" >2013<\/option><option value=\"2012\" >2012<\/option><option value=\"2011\" >2011<\/option><option value=\"2010\" >2010<\/option><option value=\"2009\" >2009<\/option><option value=\"2008\" >2008<\/option><option value=\"2007\" >2007<\/option><option value=\"2006\" >2006<\/option><option value=\"2005\" >2005<\/option><option value=\"2004\" >2004<\/option><option value=\"2002\" >2002<\/option><option value=\"2001\" >2001<\/option><option value=\"2000\" >2000<\/option><option value=\"1999\" >1999<\/option><option value=\"1998\" >1998<\/option><option value=\"1993\" >1993<\/option><option value=\"1992\" >1992<\/option><option value=\"0000\" >0000<\/option>\r\n                <\/select><select class=\"block\" title=\"All types\" name=\"type\" id=\"type\" tabindex=\"3\">\r\n                   <option value=\"\">All types<\/option>\r\n                   <option value=\"article\" >Journal Articles<\/option><option value=\"booklet\" >Booklets<\/option><option value=\"conference\" >Conferences<\/option><option value=\"inbook\" >Book Chapters<\/option><option value=\"inproceedings\" >Proceedings Articles<\/option><option value=\"mastersthesis\" >Masters Theses<\/option><option value=\"proceedings\" >Proceedings<\/option><option value=\"workshop\" >Workshops<\/option>\r\n                <\/select><select class=\"block\" title=\"All authors\" name=\"auth\" id=\"auth\" tabindex=\"5\">\r\n                   <option value=\"\">All authors<\/option>\r\n                   <option value=\"350\" > Abdi, Hamidreza<\/option><option value=\"724\" > Abdul, Sami A<\/option><option value=\"518\" > Ableson, Alan<\/option><option value=\"246\" > Abolmaesumi, P<\/option><option value=\"7\" > Abolmaesumi, Purang<\/option><option value=\"516\" > Aboofazeli, M<\/option><option value=\"489\" > Aboofazeli, Mohammad<\/option><option value=\"816\" > Abraham, M Roselle<\/option><option value=\"663\" > Abu-Omar, Nawaf<\/option><option value=\"623\" > Adams, Michael A<\/option><option value=\"604\" > Addas, Shamel<\/option><option value=\"613\" > Adebayo, Flourish Omolara<\/option><option value=\"335\" > Agarwal, Harsh<\/option><option value=\"75\" > Aghaeeaval, Mahsa<\/option><option value=\"772\" > Ahmad, Mirza Shaharyar<\/option><option value=\"723\" > Ahmadi, Negar<\/option><option value=\"522\" > Ahmidi, N<\/option><option value=\"412\" > Akbarifar, F<\/option><option value=\"82\" > Akbarifar, Faranak<\/option><option value=\"720\" > Akhtar-Danesh, Gileh-Gol<\/option><option value=\"728\" > Alaichi, Jacob A<\/option><option value=\"731\" > Alaichi, Jacob<\/option><option value=\"763\" > Alam, Armaghan<\/option><option value=\"668\" > Albacete, Samantha<\/option><option value=\"694\" > Alhabboubi, Mostafa<\/option><option value=\"706\" > Alibhai, Kameela<\/option><option value=\"699\" > Allard-Coutu, Alexandra<\/option><option value=\"153\" > Amalou, Amel<\/option><option value=\"808\" > Amhis, Nawal<\/option><option value=\"339\" > Anas, E M A<\/option><option value=\"243\" > Anas, EMA<\/option><option value=\"261\" > Anas, Emran<\/option><option value=\"267\" > Anas, Emran M A<\/option><option value=\"149\" > Anas, Emran Mohammad Abu<\/option><option value=\"49\" > Asselin, Mark<\/option><option value=\"759\" > Auer, Rebecca C<\/option><option value=\"178\" > Auyeung, Jason<\/option><option value=\"429\" > Ayukawa, I<\/option><option value=\"431\" > Ayukawa, Irene<\/option><option value=\"386\" > Aziz, S<\/option><option value=\"241\" > Azizi, S<\/option><option value=\"190\" > Azizi, Shekoofeh<\/option><option value=\"664\" > Baig, Zarrukh<\/option><option value=\"744\" > Baker, Laura<\/option><option value=\"676\" > Balvardi, Saba<\/option><option value=\"548\" > Baranzini, S<\/option><option value=\"545\" > Baranzini, S E<\/option><option value=\"368\" > Baranzini, SE<\/option><option value=\"407\" > Baranzini, Sergio<\/option><option value=\"404\" > Baranzini, Sergio E<\/option><option value=\"719\" > Barber, Evan<\/option><option value=\"211\" > Barnes, Jack A<\/option><option value=\"225\" > Barnes, Jack<\/option><option value=\"753\" > Baronas, Victoria<\/option><option value=\"31\" > Barr, Colton<\/option><option value=\"221\" > Baum, Z<\/option><option value=\"242\" > Bayat, S<\/option><option value=\"11\" > Bayat, Sharareh<\/option><option value=\"749\" > Bayat, Zubair<\/option><option value=\"498\" > Beek, M<\/option><option value=\"450\" > Behl, P<\/option><option value=\"283\" > Behnami, Delaram<\/option><option value=\"184\" > Behringer, Peter A<\/option><option value=\"76\" > Bendahan, Nathaniel<\/option><option value=\"113\" > Berman, David<\/option><option value=\"528\" > Bernard, Claude C A<\/option><option value=\"336\" > Bharat, Shyam<\/option><option value=\"16\" > Black, Peter<\/option><option value=\"506\" > Boag, A<\/option><option value=\"479\" > Boag, Alexander<\/option><option value=\"354\" > Boag, Alexander H<\/option><option value=\"493\" > Boisvert, J<\/option><option value=\"448\" > Boisvert, Jonathan<\/option><option value=\"721\" > Bondzi-Simpson, Adom<\/option><option value=\"436\" > Borschneck, D<\/option><option value=\"381\" > Borschneck, D P<\/option><option value=\"468\" > Borschneck, Dan<\/option><option value=\"402\" > Borschneck, Dan P<\/option><option value=\"224\" > Borschneck, Daniel<\/option><option value=\"311\" > Borschneck, Daniel P<\/option><option value=\"435\" > Borschneck, DP<\/option><option value=\"449\" > Borshneck, Dan<\/option><option value=\"722\" > Bowker, Riley<\/option><option value=\"790\" > Brackstone, Muriel<\/option><option value=\"691\" > Bradley, Nori<\/option><option value=\"735\" > Brennan, Kelly<\/option><option value=\"574\" > Brummendorf, T<\/option><option value=\"439\" > Burdette, E<\/option><option value=\"464\" > Burdette, E C<\/option><option value=\"466\" > Burdette, Everett C<\/option><option value=\"534\" > Caillier, Stacy J<\/option><option value=\"636\" > Camara, Mamadou<\/option><option value=\"632\" > Camara, Mamadou Samba<\/option><option value=\"756\" > Caminsky, Natasha<\/option><option value=\"487\" > Campigotto, A<\/option><option value=\"738\" > Candy, Sydney<\/option><option value=\"591\" > Carlo, Alessia Di<\/option><option value=\"734\" > Castelo, Matthew<\/option><option value=\"648\" > Caycedo-Marulanda, Antonio<\/option><option value=\"213\" > Chan, B<\/option><option value=\"150\" > Chan, Brandon<\/option><option value=\"559\" > Chan, C<\/option><option value=\"550\" > Chan, Cheryl<\/option><option value=\"401\" > Chang, S<\/option><option value=\"15\" > Chang, Silvia<\/option><option value=\"570\" > Chavez, E<\/option><option value=\"590\" > Cheesman, Joanna<\/option><option value=\"103\" > Chen, Brian<\/option><option value=\"492\" > Chen, E<\/option><option value=\"486\" > Chen, Elvis C S<\/option><option value=\"470\" > Chen, Elvis<\/option><option value=\"802\" > Chen, Ko Ta<\/option><option value=\"541\" > Chen, T K<\/option><option value=\"236\" > Cheng, Cheng-Chieh<\/option><option value=\"421\" > Chitty, David W<\/option><option value=\"547\" > Cho, C R<\/option><option value=\"786\" > Choi, Woo Jin<\/option><option value=\"775\" > Chow, Alexandra<\/option><option value=\"94\" > Chowdhury, Fahmida A<\/option><option value=\"96\" > Chowdhury, Fahmida Amin<\/option><option value=\"259\" > Choyke, P<\/option><option value=\"156\" > Choyke, Peter<\/option><option value=\"594\" > Chris, Yeung<\/option><option value=\"222\" > Chruch, Ben<\/option><option value=\"457\" > Chung, Jonathan<\/option><option value=\"310\" > Church, Ben<\/option><option value=\"809\" > Claasen, Marco<\/option><option value=\"196\" > Cobzas, D<\/option><option value=\"555\" > Comabella, Manuel<\/option><option value=\"40\" > Connolly, Laura<\/option><option value=\"209\" > Conolly, Laura<\/option><option value=\"434\" > Cordy, J<\/option><option value=\"410\" > Cordy, James<\/option><option value=\"109\" > Cotechini, Tiziana<\/option><option value=\"650\" > Curran, Jeffrey D<\/option><option value=\"789\" > Cyr, David<\/option><option value=\"703\" > D\u2019Elia, Michael A<\/option><option value=\"610\" > Damiani, Simona<\/option><option value=\"463\" > Daoud, M<\/option><option value=\"467\" > Daoud, Mohammad<\/option><option value=\"272\" > Daoud, Mohammad I<\/option><option value=\"342\" > Darras, K<\/option><option value=\"316\" > Darras, Kathryn<\/option><option value=\"514\" > Davidson, C<\/option><option value=\"507\" > Davidson, C.<\/option><option value=\"480\" > Davidson, Chris<\/option><option value=\"424\" > Day, Andrew<\/option><option value=\"672\" > Daza, Julian<\/option><option value=\"101\" > Deguet, Anton<\/option><option value=\"652\" > Deike-Hofmann, Katerina<\/option><option value=\"692\" > Deng, Shirley Xiaoxuan<\/option><option value=\"27\" > Diao, Babacar<\/option><option value=\"473\" > Dickinson, A W L<\/option><option value=\"390\" > Dickinson, Andrew W L<\/option><option value=\"247\" > Ding, Hongyi<\/option><option value=\"631\" > Diop, Idy<\/option><option value=\"685\" > Do, Uyen<\/option><option value=\"572\" > Dragowska, V<\/option><option value=\"697\" > Drung, Jeremy<\/option><option value=\"374\" > Dukelow, S<\/option><option value=\"359\" > Dukelow, S P<\/option><option value=\"299\" > Dukelow, Sean<\/option><option value=\"416\" > Dukelow, Sean P<\/option><option value=\"88\" > Dukelow, Sean P.<\/option><option value=\"387\" > Dukelow, SP<\/option><option value=\"627\" > Duncan, James<\/option><option value=\"95\" > Duncan, John S<\/option><option value=\"576\" > Duncan, T<\/option><option value=\"656\" > Effland, Alexander<\/option><option value=\"755\" > Ehlebracht, Alexa<\/option><option value=\"587\" > Ehrlich, Josh<\/option><option value=\"37\" > Ehrlich, Joshua<\/option><option value=\"765\" > El-Kefraoui, Charbel<\/option><option value=\"495\" > Eleveld, H<\/option><option value=\"526\" > Ellis, R E<\/option><option value=\"640\" > Elmi, Hanad<\/option><option value=\"22\" > Engel, Cecil Jay<\/option><option value=\"86\" > Engel, Jay<\/option><option value=\"798\" > Eom, Ashley<\/option><option value=\"14\" > Eshumani, Walid<\/option><option value=\"579\" > Fallah, A<\/option><option value=\"607\" > Fang, Alexis<\/option><option value=\"818\" > Farahmand, Mohammad<\/option><option value=\"18\" > Fauerbach, Paola Nasute<\/option><option value=\"189\" > Fedorov, Andriy<\/option><option value=\"504\" > Feldman, M<\/option><option value=\"477\" > Feldman, Michael D<\/option><option value=\"569\" > Fels, S S<\/option><option value=\"565\" > Fels, Sidney<\/option><option value=\"562\" > Fels, Sidney S<\/option><option value=\"186\" > Fennessy, Fiona M<\/option><option value=\"307\" > Fenster, A<\/option><option value=\"97\" > Fenster, Aaron<\/option><option value=\"532\" > Fernald, Guy Haskin<\/option><option value=\"382\" > Fichtinger, G<\/option><option value=\"23\" > Fichtinger, Gabor<\/option><option value=\"4\" > Fooladgar, Fahimeh<\/option><option value=\"483\" > Foroughi, Pezhman<\/option><option value=\"194\" > Frisken, S<\/option><option value=\"228\" > Frisken, Sarah<\/option><option value=\"303\" > Gaed, M<\/option><option value=\"141\" > Gaed, Mena<\/option><option value=\"581\" > Galesh-Khale, Amir Khojaste<\/option><option value=\"593\" > Gallo, Andrea<\/option><option value=\"736\" > Garfinkle, Richard<\/option><option value=\"204\" > Gasmallah, Mohammed<\/option><option value=\"677\" > Gawad, Nada<\/option><option value=\"274\" > Gerardo, Carlos D<\/option><option value=\"126\" > Gerolami, Justin<\/option><option value=\"289\" > Ghafoorian, Mohsen<\/option><option value=\"471\" > Ghanavati, Sahar<\/option><option value=\"320\" > Ghavidel, Sahar<\/option><option value=\"440\" > Ghoshal, G<\/option><option value=\"306\" > Gibson, E<\/option><option value=\"144\" > Gibson, Eli<\/option><option value=\"1\" > Gilany, Mahdi<\/option><option value=\"757\" > Gilbert, Richard<\/option><option value=\"491\" > Gill, S<\/option><option value=\"447\" > Gill, Sean<\/option><option value=\"618\" > Giovannetti, Olivia<\/option><option value=\"372\" > Glasgow, J<\/option><option value=\"415\" > Glasgow, Janice I<\/option><option value=\"389\" > Glasgow, JI<\/option><option value=\"673\" > Glass, Lyndsay T<\/option><option value=\"474\" > Gobbi, D G<\/option><option value=\"433\" > Gobbi, D<\/option><option value=\"391\" > Gobbi, David G<\/option><option value=\"409\" > Gobbi, David<\/option><option value=\"349\" > Goksel, Orcun<\/option><option value=\"201\" > Golby, A<\/option><option value=\"237\" > Golby, Alexandra<\/option><option value=\"279\" > Goldenberg, Larry<\/option><option value=\"231\" > Golland, Polina<\/option><option value=\"578\" > Golpaighani, Hashemi M R<\/option><option value=\"304\" > Gomez, J A<\/option><option value=\"385\" > Gomez, J<\/option><option value=\"142\" > Gomez, Jose A<\/option><option value=\"301\" > Goncalves, C<\/option><option value=\"312\" > Goncalves, Caroline<\/option><option value=\"469\" > Gong, Ren Hui<\/option><option value=\"175\" > Gooding, Jim<\/option><option value=\"282\" > Goyal, Apoorv<\/option><option value=\"112\" > Graham, Charles<\/option><option value=\"425\" > Gravett, Matthew<\/option><option value=\"117\" > Grebenisan, Andrew<\/option><option value=\"687\" > Greenberg, Brianna<\/option><option value=\"813\" > Greenland, Gabriela V<\/option><option value=\"625\" > Greenspan, Hayit<\/option><option value=\"546\" > Greller, L D<\/option><option value=\"556\" > Greller, Larry D<\/option><option value=\"806\" > Griffiths, Christopher<\/option><option value=\"633\" > Groves, Leah A<\/option><option value=\"28\" > Groves, Leah<\/option><option value=\"675\" > Guidolin, Keegan<\/option><option value=\"657\" > Haase, Robert<\/option><option value=\"363\" > Hacihaliloglu, I<\/option><option value=\"351\" > Hacihaliloglu, Ilker<\/option><option value=\"478\" > Hagemann, Ian<\/option><option value=\"754\" > Hajjar, Roy<\/option><option value=\"226\" > Hamidi, Soheil<\/option><option value=\"104\" > Hamilton, Alexander<\/option><option value=\"783\" > Hanna, Nader<\/option><option value=\"111\" > Hardy, Celine<\/option><option value=\"597\" > Harmanani, Mohamed<\/option><option value=\"726\" > Harrison, Lyndsay<\/option><option value=\"459\" > Hashemi, J<\/option><option value=\"355\" > Hashemi, Javad<\/option><option value=\"430\" > Hashtrudi-Zaad, K<\/option><option value=\"357\" > Hashtrudi-Zaad, Keyvan<\/option><option value=\"777\" > He, Wenjing<\/option><option value=\"441\" > Heffter, T<\/option><option value=\"679\" > Henley, Jessica<\/option><option value=\"182\" > Herz, Christian<\/option><option value=\"365\" > Himmelstein, DS<\/option><option value=\"110\" > Hindmarch, Charles<\/option><option value=\"793\" > Hirpara, Dhruvin<\/option><option value=\"33\" > Hisey, Rebecca<\/option><option value=\"747\" > Holden, James<\/option><option value=\"379\" > Holden, M<\/option><option value=\"34\" > Holden, Matthew<\/option><option value=\"704\" > Hopkins, Brent<\/option><option value=\"761\" > Hopkins, Jessica<\/option><option value=\"781\" > Hossain, Intekhab<\/option><option value=\"709\" > How, Nathan<\/option><option value=\"17\" > Hu, Zoe<\/option><option value=\"500\" > Huang, K<\/option><option value=\"714\" > Huo, Bright<\/option><option value=\"647\" > Hurlbut, David<\/option><option value=\"13\" > Hurtado, Antonio<\/option><option value=\"621\" > Iaboni, Natasha<\/option><option value=\"745\" > Ichhpuniani, Simarpreet<\/option><option value=\"277\" > Iczkowski, Kenneth A<\/option><option value=\"132\" > Idriss, Ahmedou Moulaye<\/option><option value=\"122\" > III, William M Wells<\/option><option value=\"177\" > III, William Wells M<\/option><option value=\"302\" > Imani, F<\/option><option value=\"140\" > Imani, Farhad<\/option><option value=\"680\" > Imbert, Emma<\/option><option value=\"69\" > Imtiaz, Tashifa<\/option><option value=\"614\" > Imtiaz, Uroosa<\/option><option value=\"525\" > Iordachita, I<\/option><option value=\"125\" > Iseman, Rachael<\/option><option value=\"90\" > Isen, Jonah<\/option><option value=\"513\" > Isotalo, P<\/option><option value=\"584\" > Isotalo, Phillip A<\/option><option value=\"536\" > Isotalo, Phillip<\/option><option value=\"791\" > Ivankovic, Victoria<\/option><option value=\"55\" > Izard, Jason<\/option><option value=\"24\" > Jabs, Doris<\/option><option value=\"383\" > Jaeger, M<\/option><option value=\"63\" > Jaeger, Melanie<\/option><option value=\"160\" > Jafari, Mohammad H<\/option><option value=\"135\" > Jafari, Mohammad Hossein<\/option><option value=\"462\" > Jalal, R<\/option><option value=\"444\" > Jalal, Rayhan<\/option><option value=\"124\" > Jamaspishvili, Tamara<\/option><option value=\"413\" > Jamzad, A<\/option><option value=\"3\" > Jamzad, Amoon<\/option><option value=\"115\" > Janssen, Natasja<\/option><option value=\"151\" > Janssen, Natasja N Y<\/option><option value=\"161\" > Janssen, Natasja N. Y<\/option><option value=\"162\" > Janssen, Natasja N. Y.<\/option><option value=\"9\" > Javadi, Golara<\/option><option value=\"611\" > Javed, Fiza<\/option><option value=\"804\" > Jette, Nic<\/option><option value=\"599\" > Jin, Albert<\/option><option value=\"716\" > Jogiat, Uzair<\/option><option value=\"344\" > John, P S<\/option><option value=\"375\" > John, Paul S<\/option><option value=\"318\" > John, Paul St<\/option><option value=\"361\" > John, St P<\/option><option value=\"674\" > Johnson, Garrett<\/option><option value=\"454\" > Johri, A<\/option><option value=\"456\" > Johri, A M<\/option><option value=\"420\" > Johri, Amer M<\/option><option value=\"608\" > Jomaa, Danny<\/option><option value=\"612\" > Jones, Tristan<\/option><option value=\"53\" > Joshi, Chandra<\/option><option value=\"174\" > Jutras, Jean-David<\/option><option value=\"68\" > Kaczmarek, Emily<\/option><option value=\"93\" > Kanber, Baris<\/option><option value=\"120\" > Kapur, Tina<\/option><option value=\"643\" > Kaufman, Martin<\/option><option value=\"38\" > Kaufmann, Martin<\/option><option value=\"634\" > Keita, Mohamed<\/option><option value=\"393\" > Kempston, Michael<\/option><option value=\"713\" > Keogh, Joshua<\/option><option value=\"376\" > Keri, Z<\/option><option value=\"394\" > Keri, Zsuzsanna<\/option><option value=\"136\" > Kesch, Claudia<\/option><option value=\"321\" > Khallaghi, Siavash<\/option><option value=\"800\" > Khan, Shawn<\/option><option value=\"583\" > Khankhanian, Pouya<\/option><option value=\"397\" > Khojaste, A<\/option><option value=\"322\" > Khojaste, Amir<\/option><option value=\"729\" > Kidane, Biniam<\/option><option value=\"30\" > Kikinis, Ron<\/option><option value=\"600\" > Kim, Andrew<\/option><option value=\"392\" > King, Franklin<\/option><option value=\"50\" > Kitner, Nicole<\/option><option value=\"537\" > Knott, S<\/option><option value=\"484\" > Knott, Simon<\/option><option value=\"655\" > Kobler, Erich<\/option><option value=\"51\" > Korzeniowski, Martin<\/option><option value=\"586\" > Koster, Teaghan<\/option><option value=\"779\" > Kouzmina, Ekaterina<\/option><option value=\"780\" > Koziak, Cypriana<\/option><option value=\"98\" > Kronreif, Gernot<\/option><option value=\"411\" > Kruecker, Jochen<\/option><option value=\"325\" > Kruecker, Joechen<\/option><option value=\"179\" > Kunz, Manuela<\/option><option value=\"181\" > Kuo, Alex<\/option><option value=\"560\" > Kuo, T<\/option><option value=\"330\" > Kwak, J T<\/option><option value=\"154\" > Kwak, Jin Tae<\/option><option value=\"256\" > Kwak, JT<\/option><option value=\"666\" > La, Julie<\/option><option value=\"810\" > Laboni, Natasha<\/option><option value=\"764\" > Lagace, Philip<\/option><option value=\"214\" > Laird, P<\/option><option value=\"107\" > Laird, Philip<\/option><option value=\"651\" > Laird, Philip R<\/option><option value=\"711\" > Lalande, Annie<\/option><option value=\"494\" > Lalonde, E<\/option><option value=\"458\" > Lang, Andrew<\/option><option value=\"571\" > Lansdorp, P<\/option><option value=\"568\" > Lansdorp, P M<\/option><option value=\"564\" > Lansdorp, Peter M<\/option><option value=\"696\" > Laplante, Simon<\/option><option value=\"521\" > LaPointe, A<\/option><option value=\"438\" > Lasso, A<\/option><option value=\"43\" > Lasso, Andras<\/option><option value=\"36\" > Law, Christine<\/option><option value=\"121\" > Learned-Miller, Erik<\/option><option value=\"701\" > Lee, Alex<\/option><option value=\"707\" > Lee, Christina<\/option><option value=\"32\" > Lee, Hung-Yu<\/option><option value=\"733\" > Lee, Jay<\/option><option value=\"739\" > Lee, Lawrence<\/option><option value=\"718\" > Lee, Yung<\/option><option value=\"662\" > Lena, \u00c9lise Di<\/option><option value=\"683\" > Lenet, Tori<\/option><option value=\"362\" > Lessoway, V A<\/option><option value=\"328\" > Lessoway, Victoria A<\/option><option value=\"286\" > Lessoway, Victoria<\/option><option value=\"400\" > Leveridge, M<\/option><option value=\"324\" > Leveridge, Michael<\/option><option value=\"762\" > Levin, Yosef<\/option><option value=\"582\" > Leviridge, Michael<\/option><option value=\"60\" > Lewis, Rob<\/option><option value=\"681\" > Li, Christine<\/option><option value=\"519\" > Li, K<\/option><option value=\"264\" > Li, Ming<\/option><option value=\"812\" > Li, Pengyuan<\/option><option value=\"147\" > Li, Si Jia<\/option><option value=\"281\" > Li, Sijia<\/option><option value=\"771\" > Li, Ting<\/option><option value=\"615\" > Li, Yiheng<\/option><option value=\"695\" > Lie, Jessica<\/option><option value=\"35\" > Liu, John<\/option><option value=\"180\" > Liu, Sam<\/option><option value=\"83\" > Logan, Kathryn<\/option><option value=\"79\" > Lomax, Lysa Boisse<\/option><option value=\"212\" > Loock, Hans-Peter<\/option><option value=\"580\" > Lucas, C<\/option><option value=\"278\" > Lucia, Scott M<\/option><option value=\"193\" > Luo, J<\/option><option value=\"217\" > Luo, Jie<\/option><option value=\"796\" > Luu, Shelly<\/option><option value=\"751\" > Maalouf, Michael Fares<\/option><option value=\"229\" > Machado, Ines<\/option><option value=\"732\" > Mackay, Emily<\/option><option value=\"183\" > MacNeil, Kyle<\/option><option value=\"508\" > Madabhushi, A<\/option><option value=\"481\" > Madabhushi, Anant<\/option><option value=\"406\" > Madireddy, Lohith<\/option><option value=\"605\" > Maghsoodi, Nooshin<\/option><option value=\"292\" > Mahdavi, Sara S<\/option><option value=\"769\" > Mailloux, Olivier<\/option><option value=\"766\" > Mainprize, Marguerite<\/option><option value=\"452\" > Malik, P<\/option><option value=\"423\" > Malik, Paul<\/option><option value=\"603\" > Mannina, Sophia<\/option><option value=\"700\" > Mansouri, Sarah<\/option><option value=\"46\" > March, Lucas<\/option><option value=\"577\" > Martens, U<\/option><option value=\"215\" > Maslove, D<\/option><option value=\"649\" > Maslove, David M<\/option><option value=\"106\" > Maslove, David<\/option><option value=\"451\" > Matangi, M<\/option><option value=\"422\" > Matangi, Murray<\/option><option value=\"405\" > Matsushita, Takuya<\/option><option value=\"62\" > Maxwell, Sarah<\/option><option value=\"811\" > Mayfield, Jacob<\/option><option value=\"609\" > Mayotte, Harrison<\/option><option value=\"380\" > McGraw, R<\/option><option value=\"717\" > McGuire, Anna L<\/option><option value=\"78\" > Mcinnis, Carter<\/option><option value=\"47\" > McKay, Doug<\/option><option value=\"678\" > McKechnie, Tyler<\/option><option value=\"544\" > McMichael, J<\/option><option value=\"588\" > McMullen, Julie<\/option><option value=\"152\" > Mehrtash, Alireza<\/option><option value=\"767\" > Melland-Smith, Megan<\/option><option value=\"797\" > Meloche-Dumas, L\u00e9amarie<\/option><option value=\"56\" > Menard, Alexandre<\/option><option value=\"85\" > Merchant, Shaila<\/option><option value=\"684\" > Metz, Jeff<\/option><option value=\"805\" > Mir, Zuhaib<\/option><option value=\"782\" > Mocanu, Valentin<\/option><option value=\"482\" > Moghari, Mehdi Hedjazi<\/option><option value=\"244\" > Mohamed, T<\/option><option value=\"268\" > Mohamed, Tamer<\/option><option value=\"502\" > Monaco, J P<\/option><option value=\"497\" > Monaco, James<\/option><option value=\"475\" > Monaco, James P<\/option><option value=\"558\" > Montalban, Xavier<\/option><option value=\"708\" > Moon, Jeongyoon Jenny<\/option><option value=\"333\" > Moradi, M<\/option><option value=\"505\" > Moradi, M.<\/option><option value=\"352\" > Moradi, Mehdi<\/option><option value=\"460\" > Morin, E<\/option><option value=\"356\" > Morin, Evelyn<\/option><option value=\"758\" > Morin, Genevi\u00e8ve<\/option><option value=\"294\" > Morris, William J<\/option><option value=\"169\" > Morse, Nicole<\/option><option value=\"602\" > Morton, David<\/option><option value=\"358\" > Mostafavi, S M<\/option><option value=\"538\" > Mostafavi, S<\/option><option value=\"485\" > Mostafavi, Sara<\/option><option value=\"297\" > Mostafavi, Sayyed Mostafa<\/option><option value=\"371\" > Mostafavi, SM<\/option><option value=\"446\" > Mountjoy, Katherine<\/option><option value=\"216\" > Mousavi, P<\/option><option value=\"8\" > Mousavi, Parvin<\/option><option value=\"305\" > Moussa, M<\/option><option value=\"143\" > Moussa, Madeleine<\/option><option value=\"688\" > Muaddi, Hala<\/option><option value=\"529\" > Munain, Adolfo Lopez<\/option><option value=\"620\" > Munday, Jasica<\/option><option value=\"693\" > Murphy, Patrick<\/option><option value=\"137\" > N.Y.Janssen, Natasja<\/option><option value=\"820\" > Nabi, Majid<\/option><option value=\"427\" > Nagpal, S<\/option><option value=\"369\" > Nagpal, Simrin<\/option><option value=\"300\" > Nahlawi, L<\/option><option value=\"327\" > Nahlawi, Layan Imad<\/option><option value=\"139\" > Nahlawi, Layan<\/option><option value=\"792\" > Nair, Ananya Gopika<\/option><option value=\"70\" > Nanayakkara, Jina<\/option><option value=\"535\" > Narayan, Kavitha<\/option><option value=\"606\" > Nassar, Sarah<\/option><option value=\"768\" > Nasser, Khadija<\/option><option value=\"630\" > Ndiaye, Fatou Bintou<\/option><option value=\"100\" > Ndoye, Alain<\/option><option value=\"443\" > Neubauer, P<\/option><option value=\"803\" > Newman-Bremang, Jieun<\/option><option value=\"658\" > Nezamabadi, Kasra<\/option><option value=\"795\" > Ng, Deanna<\/option><option value=\"622\" > Nicol, Christopher JB<\/option><option value=\"165\" > Nicol, Christopher<\/option><option value=\"41\" > Nikniazi, Arash<\/option><option value=\"273\" > Nir, Guy<\/option><option value=\"698\" > Nixon, Thomas<\/option><option value=\"396\" > Nouranian, S<\/option><option value=\"291\" > Nouranian, Saman<\/option><option value=\"45\" > Nunzi, Jean Michael<\/option><option value=\"119\" > O\u2019Donnell, Lauren J<\/option><option value=\"549\" > Oksenberg, J<\/option><option value=\"542\" > Oksenberg, Jorge<\/option><option value=\"530\" > Oksenberg, Jorge R<\/option><option value=\"539\" > Oksernberg, Jorge<\/option><option value=\"52\" > Olding, Tim<\/option><option value=\"646\" > Olding, Timothy<\/option><option value=\"166\" > Orr, Christine<\/option><option value=\"288\" > Osborn, Jill<\/option><option value=\"527\" > Otaegui, David<\/option><option value=\"296\" > others,<\/option><option value=\"533\" > Pachner, Andrew<\/option><option value=\"748\" > Paradis, Tiffany<\/option><option value=\"689\" > Park, Lily<\/option><option value=\"367\" > Parvin, M<\/option><option value=\"58\" > Patel, Dhruv<\/option><option value=\"725\" > Patel, Pooja<\/option><option value=\"750\" > Patel, Sunil<\/option><option value=\"715\" > Patel, Yogita S<\/option><option value=\"64\" > Patterson, Lindsay<\/option><option value=\"25\" > Perera-Ortega, Andrea<\/option><option value=\"134\" > Pesteie, Mehran<\/option><option value=\"760\" > Phang, Terry<\/option><option value=\"345\" > Pichora, D<\/option><option value=\"501\" > Pichora, D R<\/option><option value=\"319\" > Pichora, David<\/option><option value=\"540\" > Pichora, DR<\/option><option value=\"218\" > Pieper, Steve<\/option><option value=\"654\" > Pinetz, Thomas<\/option><option value=\"254\" > Pinto, P<\/option><option value=\"157\" > Pinto, Peter<\/option><option value=\"784\" > Pook, Makena<\/option><option value=\"575\" > Poon, S S S<\/option><option value=\"195\" > Popuri, K<\/option><option value=\"819\" > Pourghaderi, Ahmad Reza<\/option><option value=\"262\" > Poursartip, A<\/option><option value=\"659\" > Poursartip, Anoush<\/option><option value=\"42\" > Poushimin, Rana<\/option><option value=\"198\" > Preiswerk, F<\/option><option value=\"235\" > Preiswerk, Frank<\/option><option value=\"667\" > Purich, Kieran<\/option><option value=\"644\" > Purzner, Jamie<\/option><option value=\"645\" > Purzner, Teresa<\/option><option value=\"72\" > Pyman, Blake<\/option><option value=\"730\" > Qu, Linda Chang<\/option><option value=\"653\" > Radbruch, Alexander<\/option><option value=\"638\" > Radcliffe, Olivia<\/option><option value=\"240\" > Rajaram, A<\/option><option value=\"266\" > Rajaram, Ajay<\/option><option value=\"807\" > Rajendran, Luckshi<\/option><option value=\"395\" > Ramezani, M<\/option><option value=\"384\" > Ramezani, Mahdi<\/option><option value=\"778\" > Ramji, Karim<\/option><option value=\"340\" > Rasoulian, A<\/option><option value=\"284\" > Rasoulian, Abtin<\/option><option value=\"129\" > Ren, Kevin<\/option><option value=\"595\" > Ren, Kevin Ti Mi<\/option><option value=\"138\" > Ren, Kevin Y. M.<\/option><option value=\"67\" > Ren, Kevin Yi Mi<\/option><option value=\"48\" > Ren, Kevin YM<\/option><option value=\"71\" > Renwick, Neil<\/option><option value=\"551\" > Rio, Jordi<\/option><option value=\"206\" > Rivest, Francois<\/option><option value=\"774\" > Roach, Eileen<\/option><option value=\"740\" > Robitaille, Stephan<\/option><option value=\"585\" > Rodgers, Jessica R<\/option><option value=\"39\" > Rodgers, Jessica<\/option><option value=\"92\" > Rodionov, Roman<\/option><option value=\"814\" > Rodriguez, Sebastian<\/option><option value=\"346\" > Rohling, R<\/option><option value=\"370\" > Rohling, Robert N<\/option><option value=\"192\" > Rohling, Robert<\/option><option value=\"398\" > Romagnoli, C<\/option><option value=\"326\" > Romagnoli, Cesare<\/option><option value=\"686\" > Rouhi, Armin<\/option><option value=\"163\" > Rubino, Rachel<\/option><option value=\"642\" > Rudan, J<\/option><option value=\"148\" > Rudan, John F<\/option><option value=\"65\" > Rudan, John F.<\/option><option value=\"21\" > Rudan, John<\/option><option value=\"573\" > Rufer, N<\/option><option value=\"99\" > Ruiz-Alzola, Juan<\/option><option value=\"61\" > Ryan, Sarah GF<\/option><option value=\"741\" > Sadri, Hamid<\/option><option value=\"670\" > Safar, Ali<\/option><option value=\"280\" > Salcudean, Septimiu E<\/option><option value=\"626\" > Salcudean, Septimiu<\/option><option value=\"295\" > Salcudean, Septimu E<\/option><option value=\"509\" > Salehi, M<\/option><option value=\"517\" > Salehi, Maryam<\/option><option value=\"173\" > Salomons, Gregory<\/option><option value=\"10\" > Samadi, Samareh<\/option><option value=\"567\" > Sameti, M<\/option><option value=\"563\" > Sameti, Mohammad<\/option><option value=\"347\" > Sanchez, Antonio C<\/option><option value=\"366\" > Santaniello, A<\/option><option value=\"637\" > Santilli, Alice ML<\/option><option value=\"66\" > Santilli, Alice<\/option><option value=\"788\" > Saravana-Bawan, Bianka<\/option><option value=\"353\" > Sauerberi, Eric E<\/option><option value=\"511\" > Sauerbrei, E E<\/option><option value=\"512\" > Sauerbrei, E<\/option><option value=\"488\" > Sauerbrei, Eric E<\/option><option value=\"490\" > Sauerbrei, Eric<\/option><option value=\"671\" > Schellenberg, Morgan<\/option><option value=\"223\" > Schlenger, Christopher<\/option><option value=\"801\" > Schmitz, Erika<\/option><option value=\"176\" > Schreiner, John<\/option><option value=\"712\" > Schweitzer, Christina<\/option><option value=\"360\" > Scott, S H<\/option><option value=\"373\" > Scott, S<\/option><option value=\"388\" > Scott, SH<\/option><option value=\"417\" > Scott, Stephen H<\/option><option value=\"89\" > Scott, Stephen H.<\/option><option value=\"298\" > Scott, Stephen<\/option><option value=\"199\" > Sedghi, A<\/option><option value=\"29\" > Sedghi, Alireza<\/option><option value=\"341\" > Seitel, A<\/option><option value=\"285\" > Seitel, Alexander<\/option><option value=\"661\" > Selim, Yasser<\/option><option value=\"419\" > Sellens, Rick<\/option><option value=\"742\" > Selznick, Sydney<\/option><option value=\"414\" > Setarehdan, S K<\/option><option value=\"271\" > Setarehdan, Seyed Kamaledin<\/option><option value=\"737\" > Sharma, Sahil<\/option><option value=\"641\" > Sharp, Mackenzie<\/option><option value=\"309\" > Shatkay, H<\/option><option value=\"146\" > Shatkay, Hagit<\/option><option value=\"727\" > Shi, Ge<\/option><option value=\"77\" > Shivji, Zaitoon<\/option><option value=\"80\" > Shukla, Garima<\/option><option value=\"105\" > Sibley, Stephanie<\/option><option value=\"399\" > Siemens, D R<\/option><option value=\"114\" > Siemens, David R<\/option><option value=\"531\" > Siemens, R<\/option><option value=\"123\" > Siemens, Rob<\/option><option value=\"323\" > Siemens, Robert D<\/option><option value=\"57\" > Siemens, Robert<\/option><option value=\"455\" > Simpson, C S<\/option><option value=\"453\" > Simpson, C<\/option><option value=\"426\" > Simpson, Chris<\/option><option value=\"815\" > Simsek, Bahadir<\/option><option value=\"682\" > Skinner, Samuel<\/option><option value=\"428\" > Sojoudi, S<\/option><option value=\"12\" > Sojoudi, Samira<\/option><option value=\"589\" > Solberg, Kathryn<\/option><option value=\"817\" > Soltani, Marzieh<\/option><option value=\"543\" > Somogyi, R<\/option><option value=\"557\" > Somogyi, Roland<\/option><option value=\"293\" > Spadinger, Ingrid<\/option><option value=\"710\" > Spoyalo, Karina<\/option><option value=\"624\" > Srikanthan, Dilakshan<\/option><option value=\"705\" > Srivastava, Ananya<\/option><option value=\"552\" > Stillman, Althea<\/option><option value=\"794\" > Stockley, Cecily<\/option><option value=\"202\" > Sugiyama, M<\/option><option value=\"238\" > Sugiyama, Masashi<\/option><option value=\"773\" > Sun, Warren<\/option><option value=\"348\" > Sun, Yue<\/option><option value=\"44\" > Sunderland, Kyle<\/option><option value=\"418\" > Sutherland, Colin<\/option><option value=\"377\" > Sydor, D<\/option><option value=\"598\" > Syeda, Ayesha<\/option><option value=\"628\" > Syeda-Mahmood, Tanveer<\/option><option value=\"601\" > Szabo, Robert<\/option><option value=\"290\" > Taghipour, Mehdi<\/option><option value=\"253\" > Tahmasebi, A<\/option><option value=\"265\" > Tahmasebi, Amir<\/option><option value=\"635\" > Talla, Saidou<\/option><option value=\"799\" > Tang, Katelynn<\/option><option value=\"102\" > Taylor, Russell H.<\/option><option value=\"629\" > Taylor, Russell<\/option><option value=\"187\" > Tempany, Clare M<\/option><option value=\"219\" > Tempany, Clare<\/option><option value=\"133\" > Tfeil, Yahya<\/option><option value=\"108\" > Thirumal, Sindhura<\/option><option value=\"5\" > To, Minh Nguyen Nhat<\/option><option value=\"200\" > Toews, M<\/option><option value=\"232\" > Toews, Matthew<\/option><option value=\"185\" > Tokuda, Junichi<\/option><option value=\"617\" > Tomalty, Diane<\/option><option value=\"503\" > Tomaszewski, J E<\/option><option value=\"476\" > Tomaszewski, John E<\/option><option value=\"776\" > Trac, Jessica<\/option><option value=\"432\" > Trezise, M<\/option><option value=\"408\" > Trezise, Melissa<\/option><option value=\"188\" > Tuncali, Kemal<\/option><option value=\"258\" > Turkbey, B<\/option><option value=\"155\" > Turkbey, Baris<\/option><option value=\"332\" > Turkbey, I B<\/option><option value=\"338\" > Turkbey, Ismail Baris<\/option><option value=\"73\" > Tuschi, Thomas<\/option><option value=\"660\" > Tuschl, Thomas<\/option><option value=\"702\" > Tweedy, Jillian<\/option><option value=\"499\" > Tyryshkin, K<\/option><option value=\"74\" > Tyryshkin, Kathrin<\/option><option value=\"523\" > U-thainual, P<\/option><option value=\"233\" > Unadkat, Prashin<\/option><option value=\"378\" > Ungi, T<\/option><option value=\"20\" > Ungi, Tamas<\/option><option value=\"331\" > Uniyal, N<\/option><option value=\"337\" > Uniyal, Nishant<\/option><option value=\"669\" > Valji, Rahim<\/option><option value=\"130\" > Vanderbeck, Kaitlin<\/option><option value=\"87\" > Varma, Sonal<\/option><option value=\"118\" > Varmak, Sonal<\/option><option value=\"210\" > Vaughn, Thomas<\/option><option value=\"619\" > Velikonja, Leah<\/option><option value=\"665\" > Verhoeff, Kevin<\/option><option value=\"275\" > Vignon, Francois<\/option><option value=\"524\" > Vikal, S<\/option><option value=\"515\" > Vikal, Siddharth<\/option><option value=\"167\" > Villalobos, Dalila<\/option><option value=\"553\" > Villoslada, Pablo<\/option><option value=\"690\" > Vogt, Kelly<\/option><option value=\"91\" > Vos, Sjoerd B<\/option><option value=\"752\" > Wachtel, Alon<\/option><option value=\"592\" > Walker, G Ross<\/option><option value=\"128\" > Walker, Ross<\/option><option value=\"84\" > Wallis, Julie<\/option><option value=\"245\" > Walus, K<\/option><option value=\"269\" > Walus, Konrad<\/option><option value=\"116\" > Wang, Ami<\/option><option value=\"168\" > Wang, Jenny<\/option><option value=\"364\" > Wang, L<\/option><option value=\"403\" > Wang, Lili<\/option><option value=\"472\" > Wang, Zhen<\/option><option value=\"308\" > Ward, A<\/option><option value=\"313\" > Ward, Aaron D<\/option><option value=\"145\" > Ward, Aaron<\/option><option value=\"566\" > Ward, R K<\/option><option value=\"561\" > Ward, Rabab Kreidieh<\/option><option value=\"787\" > Watanabe, Akie<\/option><option value=\"461\" > Welch, M<\/option><option value=\"445\" > Welch, Mattea<\/option><option value=\"203\" > Wells, M<\/option><option value=\"239\" > Wells, William M<\/option><option value=\"220\" > Wells, William<\/option><option value=\"227\" > White, Phillip Jason<\/option><option value=\"770\" > Whyte, Marta<\/option><option value=\"131\" > Wiercigroch, Julia<\/option><option value=\"442\" > Williams, E<\/option><option value=\"343\" > Wilson, D<\/option><option value=\"317\" > Wilson, David<\/option><option value=\"596\" > Wilson, Paul FR<\/option><option value=\"2\" > Wilson, Paul<\/option><option value=\"276\" > Wilson, Storey<\/option><option value=\"81\" > Winston, Gavin P<\/option><option value=\"743\" > Wiseman, Vanessa<\/option><option value=\"6\" > Wodlinger, Brian<\/option><option value=\"255\" > Wood, B<\/option><option value=\"158\" > Wood, Bradford<\/option><option value=\"263\" > Woudenberg, Nathan Van<\/option><option value=\"26\" > Wu, Catherine O<\/option><option value=\"437\" > Wu, M<\/option><option value=\"465\" > Wu, Mark Z<\/option><option value=\"127\" > Wu, Victoria<\/option><option value=\"554\" > Wyatt, Matthew M<\/option><option value=\"520\" > Xiang, J<\/option><option value=\"257\" > Xu, S<\/option><option value=\"159\" > Xu, Sheng<\/option><option value=\"164\" > Yam, Scott<\/option><option value=\"252\" > Yan, P<\/option><option value=\"191\" > Yan, Pingkun<\/option><option value=\"170\" > Yates, Lauren<\/option><option value=\"287\" > Yen, David<\/option><option value=\"639\" > Yeo, Caitlin<\/option><option value=\"19\" > Yeung, Chris<\/option><option value=\"510\" > Young, P G<\/option><option value=\"496\" > Young, P<\/option><option value=\"746\" > Zhang, Lisa<\/option><option value=\"197\" > Zhang, M<\/option><option value=\"230\" > Zhang, Miaomiao<\/option><option value=\"616\" > Zhang, Richard<\/option><option value=\"234\" > Zhou, Haoyin<\/option><option value=\"54\" > Zhou, Meng<\/option><option value=\"329\" > Zhuang, B<\/option><option value=\"334\" > Zhuang, Bo<\/option><option value=\"59\" > Ziegler, Erik<\/option><option value=\"260\" > Zobeiry, Navid<\/option><option value=\"785\" > Zuckerman, Jesse<\/option><option value=\"205\" > Zulkernine, Farhana<\/option>\r\n                <\/select><div class=\"teachpress_search_button\"><input name=\"tps_button\" class=\"tp_search_button\" type=\"submit\" tabindex=\"10\" value=\"Search\"\/><\/div><\/div><\/form><div class=\"tablenav\"><div class=\"tablenav-pages\"><span class=\"displaying-num\">337 entries<\/span> <a class=\"page-numbers button disabled\">&laquo;<\/a> <a class=\"page-numbers button disabled\">&lsaquo;<\/a> 1 of 17 <a href=\"https:\/\/labs.cs.queensu.ca\/medilab\/publications\/?limit=2&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=\" title=\"next page\" class=\"page-numbers button\">&rsaquo;<\/a> <a href=\"https:\/\/labs.cs.queensu.ca\/medilab\/publications\/?limit=17&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=\" title=\"last page\" class=\"page-numbers button\">&raquo;<\/a> <\/div><\/div><div class=\"teachpress_publication_list\"><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Farahmand, Mohammad;  Jamzad, Amoon;  Fooladgar, Fahimeh;  Connolly, Laura;  Kaufmann, Martin;  Ren, Kevin Yi Mi;  Rudan, John;  McKay, Doug;  Fichtinger, Gabor;  Mousavi, Parvin<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" href=\"https:\/\/dx.doi.org\/10.1007\/s11548-025-03355-8\" title=\"FACT: Foundation Model for Assessing Cancer Tissue Margins with Mass Spectrometry\" target=\"blank\">FACT: Foundation Model for Assessing Cancer Tissue Margins with Mass Spectrometry<\/a> <span class=\"tp_pub_label_award\" title=\"Best Paper\"><i class=\"fas fa-trophy\"><\/i> Best Paper<\/span> <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">International Journal of Computer Assisted Radiology and Surgery, <\/span><span class=\"tp_pub_additional_volume\">vol. 20, <\/span><span class=\"tp_pub_additional_number\">no. 6, <\/span><span class=\"tp_pub_additional_pages\">pp. 1097-1104, <\/span><span class=\"tp_pub_additional_year\">2025<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 1861-6429<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_349\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('349','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_349\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('349','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_349\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('349','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_349\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Farahmand2025,<br \/>\r\ntitle = {FACT: Foundation Model for Assessing Cancer Tissue Margins with Mass Spectrometry},<br \/>\r\nauthor = {Mohammad Farahmand and Amoon Jamzad and Fahimeh Fooladgar and Laura Connolly and Martin Kaufmann and Kevin Yi Mi Ren and John Rudan and Doug McKay and Gabor Fichtinger and Parvin Mousavi},<br \/>\r\nurl = {https:\/\/arxiv.org\/pdf\/2504.11519},<br \/>\r\ndoi = {10.1007\/s11548-025-03355-8},<br \/>\r\nissn = {1861-6429},<br \/>\r\nyear  = {2025},<br \/>\r\ndate = {2025-06-01},<br \/>\r\nurldate = {2025-06-01},<br \/>\r\njournal = {International Journal of Computer Assisted Radiology and Surgery},<br \/>\r\nvolume = {20},<br \/>\r\nnumber = {6},<br \/>\r\npages = {1097-1104},<br \/>\r\nabstract = {Accurately classifying tissue margins during cancer surgeries is crucial for ensuring complete tumor removal. Rapid Evaporative Ionization Mass Spectrometry (REIMS), a tool for real-time intraoperative margin assessment, generates spectra that require machine learning models to support clinical decision-making. However, the scarcity of labeled data in surgical contexts presents a significant challenge. This study is the first to develop a foundation model tailored specifically for REIMS data, addressing this limitation and advancing real-time surgical margin assessment.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('349','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_349\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Accurately classifying tissue margins during cancer surgeries is crucial for ensuring complete tumor removal. Rapid Evaporative Ionization Mass Spectrometry (REIMS), a tool for real-time intraoperative margin assessment, generates spectra that require machine learning models to support clinical decision-making. However, the scarcity of labeled data in surgical contexts presents a significant challenge. This study is the first to develop a foundation model tailored specifically for REIMS data, addressing this limitation and advancing real-time surgical margin assessment.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('349','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_349\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/pdf\/2504.11519\" title=\"https:\/\/arxiv.org\/pdf\/2504.11519\" target=\"_blank\">https:\/\/arxiv.org\/pdf\/2504.11519<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1007\/s11548-025-03355-8\" title=\"Follow DOI:10.1007\/s11548-025-03355-8\" target=\"_blank\">doi:10.1007\/s11548-025-03355-8<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('349','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Fooladgar, Fahimeh;  To, Minh Nguyen Nhat;  Mousavi, Parvin;  Abolmaesumi, Purang<\/p><p class=\"tp_pub_title\">Manifold DivideMix: A semi-supervised contrastive learning framework for severe label noise <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_pages\">pp. 4012-4021, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_313\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('313','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_313\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('313','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_313\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{fooladgar2024,<br \/>\r\ntitle = {Manifold DivideMix: A semi-supervised contrastive learning framework for severe label noise},<br \/>\r\nauthor = {Fahimeh Fooladgar and Minh Nguyen Nhat To and Parvin Mousavi and Purang Abolmaesumi},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\npages = {4012-4021},<br \/>\r\nabstract = {Deep neural networks have proven to be highly effective when large amounts of data with clean labels are available. However their performance degrades when training data contains noisy labels leading to poor generalization on the test set. Real-world datasets contain noisy label samples that either have similar visual semantics to other classes (in-distribution) or have no semantic relevance to any class (out-of-distribution) in the dataset. Most state-of-the-art methods leverage ID labeled noisy samples as unlabeled data for semi-supervised learning but OOD labeled noisy samples cannot be used in this way because they do not belong to any class within the dataset. Hence in this paper we propose incorporating the information from all the training data by leveraging the benefits of self-supervised training. Our method aims to extract a meaningful and generalizable embedding space for each sample regardless of its label. Then we employ a simple yet effective K-nearest neighbor method to remove portions of out-of-distribution samples. By discarding these samples we propose an iterative\" Manifold DivideMix\" algorithm to find clean and noisy samples and train our model in a semi-supervised way. In addition we propose\" MixEMatch\" a new algorithm for the semi-supervised step that involves mixup augmentation at the input and final hidden representations of the model. This will extract better representations by interpolating both in the input and manifold spaces. Extensive experiments on multiple synthetic-noise image benchmarks and real-world web-crawled datasets demonstrate the effectiveness of our proposed framework. Code is \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('313','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_313\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Deep neural networks have proven to be highly effective when large amounts of data with clean labels are available. However their performance degrades when training data contains noisy labels leading to poor generalization on the test set. Real-world datasets contain noisy label samples that either have similar visual semantics to other classes (in-distribution) or have no semantic relevance to any class (out-of-distribution) in the dataset. Most state-of-the-art methods leverage ID labeled noisy samples as unlabeled data for semi-supervised learning but OOD labeled noisy samples cannot be used in this way because they do not belong to any class within the dataset. Hence in this paper we propose incorporating the information from all the training data by leveraging the benefits of self-supervised training. Our method aims to extract a meaningful and generalizable embedding space for each sample regardless of its label. Then we employ a simple yet effective K-nearest neighbor method to remove portions of out-of-distribution samples. By discarding these samples we propose an iterative\" Manifold DivideMix\" algorithm to find clean and noisy samples and train our model in a semi-supervised way. In addition we propose\" MixEMatch\" a new algorithm for the semi-supervised step that involves mixup augmentation at the input and final hidden representations of the model. This will extract better representations by interpolating both in the input and manifold spaces. Extensive experiments on multiple synthetic-noise image benchmarks and real-world web-crawled datasets demonstrate the effectiveness of our proposed framework. Code is \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('313','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Mannina, Sophia;  Addas, Shamel;  Abolmaesumi, Purang;  Mousavi, Parvin;  Maghsoodi, Nooshin;  Nassar, Sarah;  Maslove, David<\/p><p class=\"tp_pub_title\">Digital Nudges in Healthcare Contexts: An Information Systems Perspective <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_312\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('312','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_312\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('312','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_312\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{mannina2024b,<br \/>\r\ntitle = {Digital Nudges in Healthcare Contexts: An Information Systems Perspective},<br \/>\r\nauthor = {Sophia Mannina and Shamel Addas and Purang Abolmaesumi and Parvin Mousavi and Nooshin Maghsoodi and Sarah Nassar and David Maslove},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\nabstract = {Digital transformation has presented healthcare providers with new tools, roles, and challenges related to patient care. Although digital technologies like electronic health records can offer valuable information concerning patients' needs, the growing volume of data that healthcare providers receive through these tools can contribute to information overload and alert fatigue. Nudging is a behavioural economics technique that can be applied to guide healthcare providers toward optimal care decisions while limiting information overload. To better understand the application of this technique, we perform a systematic literature review that explores digital nudges oriented toward healthcare providers from an information systems perspective. This review identifies positive and negative outcomes of digital nudges and presents design principles that can guide development of nudges directed toward healthcare providers. Opportunities are discussed to further assess digital nudges through the information systems lens.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('312','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_312\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Digital transformation has presented healthcare providers with new tools, roles, and challenges related to patient care. Although digital technologies like electronic health records can offer valuable information concerning patients' needs, the growing volume of data that healthcare providers receive through these tools can contribute to information overload and alert fatigue. Nudging is a behavioural economics technique that can be applied to guide healthcare providers toward optimal care decisions while limiting information overload. To better understand the application of this technique, we perform a systematic literature review that explores digital nudges oriented toward healthcare providers from an information systems perspective. This review identifies positive and negative outcomes of digital nudges and presents design principles that can guide development of nudges directed toward healthcare providers. Opportunities are discussed to further assess digital nudges through the information systems lens.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('312','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Harmanani, Mohamed;  Wilson, Paul FR;  Fooladgar, Fahimeh;  Jamzad, Amoon;  Gilany, Mahdi;  To, Minh Nguyen Nhat;  Wodlinger, Brian;  Abolmaesumi, Purang;  Mousavi, Parvin<\/p><p class=\"tp_pub_title\">Benchmarking image transformers for prostate cancer detection from ultrasound data <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_pages\">pp. 245-251, <\/span><span class=\"tp_pub_additional_publisher\">SPIE, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_310\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('310','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_310\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('310','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_310\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{harmanani2024b,<br \/>\r\ntitle = {Benchmarking image transformers for prostate cancer detection from ultrasound data},<br \/>\r\nauthor = {Mohamed Harmanani and Paul FR Wilson and Fahimeh Fooladgar and Amoon Jamzad and Mahdi Gilany and Minh Nguyen Nhat To and Brian Wodlinger and Purang Abolmaesumi and Parvin Mousavi},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\nvolume = {12928},<br \/>\r\npages = {245-251},<br \/>\r\npublisher = {SPIE},<br \/>\r\nabstract = {PURPOSE <br \/>\r\nDeep learning methods for classifying prostate cancer (PCa) in ultrasound images typically employ convolutional neural networks (CNN) to detect cancer in small regions of interest (ROI) along a needle trace region. However, this approach suffers from weak labelling, since the ground-truth histopathology labels do not describe the properties of individual ROIs. Recently, multi-scale approaches have sought to mitigate this issue by combining the context awareness of transformers with a convolutional feature extractor to detect cancer from multiple ROIs using multiple-instance learning (MIL). In this work, we present a detailed study of several image transformer architectures for both ROI-scale and multi-scale classification, and a comparison of the performance of CNNs and transformers for ultrasound-based prostate cancer classification. We also design a novel multi-objective learning strategy that \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('310','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_310\" style=\"display:none;\"><div class=\"tp_abstract_entry\">PURPOSE <br \/>\r\nDeep learning methods for classifying prostate cancer (PCa) in ultrasound images typically employ convolutional neural networks (CNN) to detect cancer in small regions of interest (ROI) along a needle trace region. However, this approach suffers from weak labelling, since the ground-truth histopathology labels do not describe the properties of individual ROIs. Recently, multi-scale approaches have sought to mitigate this issue by combining the context awareness of transformers with a convolutional feature extractor to detect cancer from multiple ROIs using multiple-instance learning (MIL). In this work, we present a detailed study of several image transformer architectures for both ROI-scale and multi-scale classification, and a comparison of the performance of CNNs and transformers for ultrasound-based prostate cancer classification. We also design a novel multi-objective learning strategy that \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('310','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> To, Minh Nguyen Nhat;  Fooladgar, Fahimeh;  Wilson, Paul;  Harmanani, Mohamed;  Gilany, Mahdi;  Sojoudi, Samira;  Jamzad, Amoon;  Chang, Silvia;  Black, Peter;  Mousavi, Parvin;  Abolmaesumi, Purang<\/p><p class=\"tp_pub_title\">LensePro: Label noise-tolerant prototype-based network for improving cancer detection in prostate ultrasound with limited annotations <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">International Journal of Computer Assisted Radiology and Surgery, <\/span><span class=\"tp_pub_additional_pages\">pp. 1-8, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_307\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('307','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_307\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('307','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_307\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{to2024b,<br \/>\r\ntitle = {LensePro: Label noise-tolerant prototype-based network for improving cancer detection in prostate ultrasound with limited annotations},<br \/>\r\nauthor = {Minh Nguyen Nhat To and Fahimeh Fooladgar and Paul Wilson and Mohamed Harmanani and Mahdi Gilany and Samira Sojoudi and Amoon Jamzad and Silvia Chang and Peter Black and Parvin Mousavi and Purang Abolmaesumi},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\njournal = {International Journal of Computer Assisted Radiology and Surgery},<br \/>\r\npages = {1-8},<br \/>\r\npublisher = {Springer International Publishing},<br \/>\r\nabstract = {Purpose <br \/>\r\nThe standard of care for prostate cancer (PCa) diagnosis is the histopathological analysis of tissue samples obtained via transrectal ultrasound (TRUS) guided biopsy. Models built with deep neural networks (DNNs) hold the potential for direct PCa detection from TRUS, which allows targeted biopsy and subsequently enhances outcomes. Yet, there are ongoing challenges with training robust models, stemming from issues such as noisy labels, out-of-distribution (OOD) data, and limited labeled data. <br \/>\r\nMethods <br \/>\r\nThis study presents LensePro, a unified method that not only excels in label efficiency but also demonstrates robustness against label noise and OOD data. LensePro comprises two key stages: first, self-supervised learning to extract high-quality feature representations from abundant unlabeled TRUS data and, second, label noise-tolerant prototype-based learning to classify the extracted features \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('307','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_307\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Purpose <br \/>\r\nThe standard of care for prostate cancer (PCa) diagnosis is the histopathological analysis of tissue samples obtained via transrectal ultrasound (TRUS) guided biopsy. Models built with deep neural networks (DNNs) hold the potential for direct PCa detection from TRUS, which allows targeted biopsy and subsequently enhances outcomes. Yet, there are ongoing challenges with training robust models, stemming from issues such as noisy labels, out-of-distribution (OOD) data, and limited labeled data. <br \/>\r\nMethods <br \/>\r\nThis study presents LensePro, a unified method that not only excels in label efficiency but also demonstrates robustness against label noise and OOD data. LensePro comprises two key stages: first, self-supervised learning to extract high-quality feature representations from abundant unlabeled TRUS data and, second, label noise-tolerant prototype-based learning to classify the extracted features \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('307','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Connolly, Laura;  Fooladgar, Fahimeh;  Jamzad, Amoon;  Kaufmann, Martin;  Syeda, Ayesha;  Ren, Kevin;  Abolmaesumi, Purang;  Rudan, John F;  McKay, Doug;  Fichtinger, Gabor;  Mousavi, Parvin<\/p><p class=\"tp_pub_title\">ImSpect: Image-driven self-supervised learning for surgical margin evaluation with mass spectrometry <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">International Journal of Computer Assisted Radiology and Surgery, <\/span><span class=\"tp_pub_additional_pages\">pp. 1-8, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_306\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('306','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_306\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('306','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_306\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{connolly2024b,<br \/>\r\ntitle = {ImSpect: Image-driven self-supervised learning for surgical margin evaluation with mass spectrometry},<br \/>\r\nauthor = {Laura Connolly and Fahimeh Fooladgar and Amoon Jamzad and Martin Kaufmann and Ayesha Syeda and Kevin Ren and Purang Abolmaesumi and John F Rudan and Doug McKay and Gabor Fichtinger and Parvin Mousavi},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\njournal = {International Journal of Computer Assisted Radiology and Surgery},<br \/>\r\npages = {1-8},<br \/>\r\npublisher = {Springer International Publishing},<br \/>\r\nabstract = {Purpose <br \/>\r\nReal-time assessment of surgical margins is critical for favorable outcomes in cancer patients. The iKnife is a mass spectrometry device that has demonstrated potential for margin detection in cancer surgery. Previous studies have shown that using deep learning on iKnife data can facilitate real-time tissue characterization. However, none of the existing literature on the iKnife facilitate the use of publicly available, state-of-the-art pretrained networks or datasets that have been used in computer vision and other domains. <br \/>\r\nMethods <br \/>\r\nIn a new framework we call ImSpect, we convert 1D iKnife data, captured during basal cell carcinoma (BCC) surgery, into 2D images in order to capitalize on state-of-the-art image classification networks. We also use self-supervision to leverage large amounts of unlabeled, intraoperative data to accommodate the data requirements of these networks. <br \/>\r\nResults <br \/>\r\nThrough extensive ablation \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('306','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_306\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Purpose <br \/>\r\nReal-time assessment of surgical margins is critical for favorable outcomes in cancer patients. The iKnife is a mass spectrometry device that has demonstrated potential for margin detection in cancer surgery. Previous studies have shown that using deep learning on iKnife data can facilitate real-time tissue characterization. However, none of the existing literature on the iKnife facilitate the use of publicly available, state-of-the-art pretrained networks or datasets that have been used in computer vision and other domains. <br \/>\r\nMethods <br \/>\r\nIn a new framework we call ImSpect, we convert 1D iKnife data, captured during basal cell carcinoma (BCC) surgery, into 2D images in order to capitalize on state-of-the-art image classification networks. We also use self-supervision to leverage large amounts of unlabeled, intraoperative data to accommodate the data requirements of these networks. <br \/>\r\nResults <br \/>\r\nThrough extensive ablation \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('306','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Yeung, Chris;  Ungi, Tamas;  Hu, Zoe;  Jamzad, Amoon;  Kaufmann, Martin;  Walker, Ross;  Merchant, Shaila;  Engel, Cecil Jay;  Jabs, Doris;  Rudan, John;  Mousavi, Parvin;  Fichtinger, Gabor<\/p><p class=\"tp_pub_title\">From quantitative metrics to clinical success: assessing the utility of deep learning for tumor segmentation in breast surgery <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">International Journal of Computer Assisted Radiology and Surgery, <\/span><span class=\"tp_pub_additional_pages\">pp. 1-9, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_305\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('305','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_305\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('305','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_305\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{yeung2024b,<br \/>\r\ntitle = {From quantitative metrics to clinical success: assessing the utility of deep learning for tumor segmentation in breast surgery},<br \/>\r\nauthor = {Chris Yeung and Tamas Ungi and Zoe Hu and Amoon Jamzad and Martin Kaufmann and Ross Walker and Shaila Merchant and Cecil Jay Engel and Doris Jabs and John Rudan and Parvin Mousavi and Gabor Fichtinger},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\njournal = {International Journal of Computer Assisted Radiology and Surgery},<br \/>\r\npages = {1-9},<br \/>\r\npublisher = {Springer International Publishing},<br \/>\r\nabstract = {Purpose <br \/>\r\nPreventing positive margins is essential for ensuring favorable patient outcomes following breast-conserving surgery (BCS). Deep learning has the potential to enable this by automatically contouring the tumor and guiding resection in real time. However, evaluation of such models with respect to pathology outcomes is necessary for their successful translation into clinical practice. <br \/>\r\nMethods <br \/>\r\nSixteen deep learning models based on established architectures in the literature are trained on 7318 ultrasound images from 33 patients. Models are ranked by an expert based on their contours generated from images in our test set. Generated contours from each model are also analyzed using recorded cautery trajectories of five navigated BCS cases to predict margin status. Predicted margins are compared with pathology reports. <br \/>\r\nResults <br \/>\r\nThe best-performing model using both quantitative evaluation and our visual \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('305','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_305\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Purpose <br \/>\r\nPreventing positive margins is essential for ensuring favorable patient outcomes following breast-conserving surgery (BCS). Deep learning has the potential to enable this by automatically contouring the tumor and guiding resection in real time. However, evaluation of such models with respect to pathology outcomes is necessary for their successful translation into clinical practice. <br \/>\r\nMethods <br \/>\r\nSixteen deep learning models based on established architectures in the literature are trained on 7318 ultrasound images from 33 patients. Models are ranked by an expert based on their contours generated from images in our test set. Generated contours from each model are also analyzed using recorded cautery trajectories of five navigated BCS cases to predict margin status. Predicted margins are compared with pathology reports. <br \/>\r\nResults <br \/>\r\nThe best-performing model using both quantitative evaluation and our visual \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('305','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Kaufmann, Martin;  Jamzad, Amoon;  Ungi, Tamas;  Rodgers, Jessica R;  Koster, Teaghan;  Yeung, Chris;  Ehrlich, Josh;  Santilli, Alice;  Asselin, Mark;  Janssen, Natasja;  McMullen, Julie;  Solberg, Kathryn;  Cheesman, Joanna;  Carlo, Alessia Di;  Ren, Kevin Yi Mi;  Varma, Sonal;  Merchant, Shaila;  Engel, Cecil Jay;  Walker, G Ross;  Gallo, Andrea;  Jabs, Doris;  Mousavi, Parvin;  Fichtinger, Gabor;  Rudan, John F<\/p><p class=\"tp_pub_title\">Abstract PO2-23-07: Three-dimensional navigated mass spectrometry for intraoperative margin assessment during breast cancer surgery <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">Cancer Research, <\/span><span class=\"tp_pub_additional_volume\">vol. 84, <\/span><span class=\"tp_pub_additional_number\">no. 9_Supplement, <\/span><span class=\"tp_pub_additional_pages\">pp. PO2-23-07-PO2-23-07, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_304\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('304','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_304\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('304','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_304\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{kaufmann2024c,<br \/>\r\ntitle = {Abstract PO2-23-07: Three-dimensional navigated mass spectrometry for intraoperative margin assessment during breast cancer surgery},<br \/>\r\nauthor = {Martin Kaufmann and Amoon Jamzad and Tamas Ungi and Jessica R Rodgers and Teaghan Koster and Chris Yeung and Josh Ehrlich and Alice Santilli and Mark Asselin and Natasja Janssen and Julie McMullen and Kathryn Solberg and Joanna Cheesman and Alessia Di Carlo and Kevin Yi Mi Ren and Sonal Varma and Shaila Merchant and Cecil Jay Engel and G Ross Walker and Andrea Gallo and Doris Jabs and Parvin Mousavi and Gabor Fichtinger and John F Rudan},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\njournal = {Cancer Research},<br \/>\r\nvolume = {84},<br \/>\r\nnumber = {9_Supplement},<br \/>\r\npages = {PO2-23-07-PO2-23-07},<br \/>\r\npublisher = {The American Association for Cancer Research},<br \/>\r\nabstract = {Positive resection margins occur in approximately 25% of breast cancer (BCa) surgeries, requiring re-operation. Margin status is not routinely available during surgery; thus, technologies that identify residual cancer on the specimen or cavity are needed to provide intraoperative decision support that may reduce positive margin rates. Rapid evaporative ionization mass spectrometry (REIMS) is an emerging technique that chemically profiles the plume generated by tissue cauterization to classify the ablated tissue as either cancerous or non-cancerous, on the basis of detected lipid species. Although REIMS can distinguish cancer and non-cancerous breast tissue by the signals generated, it does not indicate the location of the classified tissue in real-time. Our objective was to combine REIMS with spatio-temporal navigation (navigated REIMS), and to compare performance of navigated REIMS with conventional \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('304','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_304\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Positive resection margins occur in approximately 25% of breast cancer (BCa) surgeries, requiring re-operation. Margin status is not routinely available during surgery; thus, technologies that identify residual cancer on the specimen or cavity are needed to provide intraoperative decision support that may reduce positive margin rates. Rapid evaporative ionization mass spectrometry (REIMS) is an emerging technique that chemically profiles the plume generated by tissue cauterization to classify the ablated tissue as either cancerous or non-cancerous, on the basis of detected lipid species. Although REIMS can distinguish cancer and non-cancerous breast tissue by the signals generated, it does not indicate the location of the classified tissue in real-time. Our objective was to combine REIMS with spatio-temporal navigation (navigated REIMS), and to compare performance of navigated REIMS with conventional \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('304','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Kaufmann, Martin;  Jamzad, Amoon;  Ungi, Tamas;  Rodgers, Jessica;  Koster, Teaghan;  Chris, Yeung;  Janssen, Natasja;  McMullen, Julie;  Solberg, Kathryn;  Cheesman, Joanna;  Ren, Kevin Ti Mi;  Varma, Sonal;  Merchant, Shaila;  Engel, Cecil Jay;  Walker, G Ross;  Gallo, Andrea;  Jabs, Doris;  Mousavi, Parvin;  Fichtinger, Gabor;  Rudan, John<\/p><p class=\"tp_pub_title\">Three-dimensional navigated mass spectrometry for intraoperative margin assessment during breast cancer surgery <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_volume\">vol. 31, <\/span><span class=\"tp_pub_additional_number\">no. 1, <\/span><span class=\"tp_pub_additional_pages\">pp. S10-S10, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_301\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('301','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_301\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{kaufmann2024b,<br \/>\r\ntitle = {Three-dimensional navigated mass spectrometry for intraoperative margin assessment during breast cancer surgery},<br \/>\r\nauthor = {Martin Kaufmann and Amoon Jamzad and Tamas Ungi and Jessica Rodgers and Teaghan Koster and Yeung Chris and Natasja Janssen and Julie McMullen and Kathryn Solberg and Joanna Cheesman and Kevin Ti Mi Ren and Sonal Varma and Shaila Merchant and Cecil Jay Engel and G Ross Walker and Andrea Gallo and Doris Jabs and Parvin Mousavi and Gabor Fichtinger and John Rudan},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\nvolume = {31},<br \/>\r\nnumber = {1},<br \/>\r\npages = {S10-S10},<br \/>\r\npublisher = {SPRINGER},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('301','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Kim, Andrew;  Yeung, Chris;  Szabo, Robert;  Sunderland, Kyle;  Hisey, Rebecca;  Morton, David;  Kikinis, Ron;  Diao, Babacar;  Mousavi, Parvin;  Ungi, Tamas;  Fichtinger, Gabor<\/p><p class=\"tp_pub_title\">Percutaneous nephrostomy needle guidance using real-time 3D anatomical visualization with live ultrasound segmentation <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_pages\">pp. 163-168, <\/span><span class=\"tp_pub_additional_publisher\">SPIE, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_299\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('299','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_299\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('299','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_299\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{kim2024,<br \/>\r\ntitle = {Percutaneous nephrostomy needle guidance using real-time 3D anatomical visualization with live ultrasound segmentation},<br \/>\r\nauthor = {Andrew Kim and Chris Yeung and Robert Szabo and Kyle Sunderland and Rebecca Hisey and David Morton and Ron Kikinis and Babacar Diao and Parvin Mousavi and Tamas Ungi and Gabor Fichtinger},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\nvolume = {12928},<br \/>\r\npages = {163-168},<br \/>\r\npublisher = {SPIE},<br \/>\r\nabstract = {PURPOSE <br \/>\r\nPercutaneous nephrostomy is a commonly performed procedure to drain urine to provide relief in patients with hydronephrosis. Conventional percutaneous nephrostomy needle guidance methods can be difficult, expensive, or not portable. We propose an open-source real-time 3D anatomical visualization aid for needle guidance with live ultrasound segmentation and 3D volume reconstruction using free, open-source software. <br \/>\r\nMETHODS <br \/>\r\nBasic hydronephrotic kidney phantoms were created, and recordings of these models were manually segmented and used to train a deep learning model that makes live segmentation predictions to perform live 3D volume reconstruction of the fluid-filled cavity. Participants performed 5 needle insertions with the visualization aid and 5 insertions with ultrasound needle guidance on a kidney phantom in randomized order, and these were recorded. Recordings of the \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('299','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_299\" style=\"display:none;\"><div class=\"tp_abstract_entry\">PURPOSE <br \/>\r\nPercutaneous nephrostomy is a commonly performed procedure to drain urine to provide relief in patients with hydronephrosis. Conventional percutaneous nephrostomy needle guidance methods can be difficult, expensive, or not portable. We propose an open-source real-time 3D anatomical visualization aid for needle guidance with live ultrasound segmentation and 3D volume reconstruction using free, open-source software. <br \/>\r\nMETHODS <br \/>\r\nBasic hydronephrotic kidney phantoms were created, and recordings of these models were manually segmented and used to train a deep learning model that makes live segmentation predictions to perform live 3D volume reconstruction of the fluid-filled cavity. Participants performed 5 needle insertions with the visualization aid and 5 insertions with ultrasound needle guidance on a kidney phantom in randomized order, and these were recorded. Recordings of the \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('299','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Akbarifar, Faranak;  Dukelow, Sean P;  Jin, Albert;  Mousavi, Parvin;  Scott, Stephen H<\/p><p class=\"tp_pub_title\">Optimizing Stroke Detection Using Evidential Networks and Uncertainty-Based Refinement <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_298\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('298','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_298\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('298','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_298\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{akbarifar2024,<br \/>\r\ntitle = {Optimizing Stroke Detection Using Evidential Networks and Uncertainty-Based Refinement},<br \/>\r\nauthor = {Faranak Akbarifar and Sean P Dukelow and Albert Jin and Parvin Mousavi and Stephen H Scott},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\nabstract = {Background: <br \/>\r\nTechnologies such as interactive robotics and motion capture systems permit the development of kinematic-based approaches to assess motor impairments in stroke survivors. Here we utilise the Kinarm Exoskeleton robotic system and deep learning techniques to explore differences in motor performance between healthy controls, individuals with stroke and transient ischemic attacks (TIA). <br \/>\r\nMethods: <br \/>\r\nBuilding upon previous research that employed deep learning methods to distinguish between minimally impaired stroke patients and healthy controls using Kinarm data, this study introduces a novel dimension by estimating the confidence or uncertainty of the model's predictions. An evidential network is employed to measure this confidence, which subsequently aids in the refinement of training and testing datasets. <br \/>\r\nResults: <br \/>\r\nThe application of deep learning techniques in this context proves to be promising \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('298','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_298\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Background: <br \/>\r\nTechnologies such as interactive robotics and motion capture systems permit the development of kinematic-based approaches to assess motor impairments in stroke survivors. Here we utilise the Kinarm Exoskeleton robotic system and deep learning techniques to explore differences in motor performance between healthy controls, individuals with stroke and transient ischemic attacks (TIA). <br \/>\r\nMethods: <br \/>\r\nBuilding upon previous research that employed deep learning methods to distinguish between minimally impaired stroke patients and healthy controls using Kinarm data, this study introduces a novel dimension by estimating the confidence or uncertainty of the model's predictions. An evidential network is employed to measure this confidence, which subsequently aids in the refinement of training and testing datasets. <br \/>\r\nResults: <br \/>\r\nThe application of deep learning techniques in this context proves to be promising \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('298','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Wilson, Paul FR;  Harmanani, Mohamed;  To, Minh Nguyen Nhat;  Gilany, Mahdi;  Jamzad, Amoon;  Fooladgar, Fahimeh;  Wodlinger, Brian;  Abolmaesumi, Purang;  Mousavi, Parvin<\/p><p class=\"tp_pub_title\">Toward confident prostate cancer detection using ultrasound: a multi-center study <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">International Journal of Computer Assisted Radiology and Surgery, <\/span><span class=\"tp_pub_additional_pages\">pp. 1-9, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_293\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('293','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_293\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('293','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_293\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{wilson2024,<br \/>\r\ntitle = {Toward confident prostate cancer detection using ultrasound: a multi-center study},<br \/>\r\nauthor = {Paul FR Wilson and Mohamed Harmanani and Minh Nguyen Nhat To and Mahdi Gilany and Amoon Jamzad and Fahimeh Fooladgar and Brian Wodlinger and Purang Abolmaesumi and Parvin Mousavi},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\njournal = {International Journal of Computer Assisted Radiology and Surgery},<br \/>\r\npages = {1-9},<br \/>\r\npublisher = {Springer International Publishing},<br \/>\r\nabstract = {Purpose <br \/>\r\nDeep learning-based analysis of micro-ultrasound images to detect cancerous lesions is a promising tool for improving prostate cancer (PCa) diagnosis. An ideal model should confidently identify cancer while responding with appropriate uncertainty when presented with out-of-distribution inputs that arise during deployment due to imaging artifacts and the biological heterogeneity of patients and prostatic tissue. <br \/>\r\nMethods <br \/>\r\nUsing micro-ultrasound data from 693 patients across 5 clinical centers who underwent micro-ultrasound guided prostate biopsy, we train and evaluate convolutional neural network models for PCa detection. To improve robustness to out-of-distribution inputs, we employ and comprehensively benchmark several state-of-the-art uncertainty estimation methods. <br \/>\r\nResults <br \/>\r\nPCa detection models achieve performance scores up to average AUROC with a 10-fold cross validation setup. Models with \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('293','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_293\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Purpose <br \/>\r\nDeep learning-based analysis of micro-ultrasound images to detect cancerous lesions is a promising tool for improving prostate cancer (PCa) diagnosis. An ideal model should confidently identify cancer while responding with appropriate uncertainty when presented with out-of-distribution inputs that arise during deployment due to imaging artifacts and the biological heterogeneity of patients and prostatic tissue. <br \/>\r\nMethods <br \/>\r\nUsing micro-ultrasound data from 693 patients across 5 clinical centers who underwent micro-ultrasound guided prostate biopsy, we train and evaluate convolutional neural network models for PCa detection. To improve robustness to out-of-distribution inputs, we employ and comprehensively benchmark several state-of-the-art uncertainty estimation methods. <br \/>\r\nResults <br \/>\r\nPCa detection models achieve performance scores up to average AUROC with a 10-fold cross validation setup. Models with \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('293','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_mastersthesis\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Farahmand, Mohammad<\/p><p class=\"tp_pub_title\">End-to-End Object Tracking with Spatio-Temporal Transformers <span class=\"tp_pub_type tp_  mastersthesis\">Masters Thesis<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_school\">Iran University of Science and Technology, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_348\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('348','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_348\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@mastersthesis{farahmand2023end,<br \/>\r\ntitle = {End-to-End Object Tracking with Spatio-Temporal Transformers},<br \/>\r\nauthor = {Mohammad Farahmand},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-06-01},<br \/>\r\nschool = {Iran University of Science and Technology},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {mastersthesis}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('348','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Bayat, Sharareh;  Jamzad, Amoon;  Zobeiry, Navid;  Poursartip, Anoush;  Mousavi, Parvin;  Abolmaesumi, Purang<\/p><p class=\"tp_pub_title\">Temporal enhanced Ultrasound: A new method for detection of porosity defects in composites <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">Composites Part A: Applied Science and Manufacturing, <\/span><span class=\"tp_pub_additional_volume\">vol. 164, <\/span><span class=\"tp_pub_additional_pages\">pp. 107259, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_338\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('338','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_338\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('338','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_338\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{bayat2023,<br \/>\r\ntitle = {Temporal enhanced Ultrasound: A new method for detection of porosity defects in composites},<br \/>\r\nauthor = {Sharareh Bayat and Amoon Jamzad and Navid Zobeiry and Anoush Poursartip and Parvin Mousavi and Purang Abolmaesumi},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-01-01},<br \/>\r\njournal = {Composites Part A: Applied Science and Manufacturing},<br \/>\r\nvolume = {164},<br \/>\r\npages = {107259},<br \/>\r\npublisher = {Elsevier},<br \/>\r\nabstract = {Non-Destructive Evaluation (NDE) methods are commonly employed for identifying porosity, which is one of the most common manufacturing defects observed in composite structures. Among current widely used approaches are conventional ultrasonic methods such as pulse-echo analysis based on loss of signal amplitude. Application of these conventional ultrasonic methods, however, can be challenging in cases where the loss of signal is negligible, such as with porosity. In this paper, we propose Temporal-enhanced Ultrasound (TeUS) as a novel ultrasound-based imaging technique for NDE of composites. TeUS represents the analysis of a sequence of ultrasound images obtained from composites by varying an image acquisition parameter, such as the focal point, over the sequence. We present details on the analytical formulation of TeUS, followed by extensive simulation and experimental results to \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('338','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_338\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Non-Destructive Evaluation (NDE) methods are commonly employed for identifying porosity, which is one of the most common manufacturing defects observed in composite structures. Among current widely used approaches are conventional ultrasonic methods such as pulse-echo analysis based on loss of signal amplitude. Application of these conventional ultrasonic methods, however, can be challenging in cases where the loss of signal is negligible, such as with porosity. In this paper, we propose Temporal-enhanced Ultrasound (TeUS) as a novel ultrasound-based imaging technique for NDE of composites. TeUS represents the analysis of a sequence of ultrasound images obtained from composites by varying an image acquisition parameter, such as the focal point, over the sequence. We present details on the analytical formulation of TeUS, followed by extensive simulation and experimental results to \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('338','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Nezamabadi, Kasra<\/p><p class=\"tp_pub_title\">XplainScar: Explainable Artificial Intelligence to Identify and Localize Left Ventricular Scar in Hypertrophic Cardiomyopathy from 12-lead Electrocardiogram <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_337\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('337','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_337\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('337','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_337\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{nezamabadi2023,<br \/>\r\ntitle = {XplainScar: Explainable Artificial Intelligence to Identify and Localize Left Ventricular Scar in Hypertrophic Cardiomyopathy from 12-lead Electrocardiogram},<br \/>\r\nauthor = {Kasra Nezamabadi},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-01-01},<br \/>\r\nabstract = {Myocardial scar in the left ventricular contributes significantly to sudden cardiac death in hypertrophic cardiomyopathy (HCM). Although late gadolinium-contrast enhanced magnetic resonance imaging is commonly used for detecting HCM scar, its high cost, limited availability, and susceptibility to artifacts from implanted devices make it unsuitable for ongoing scar progression and risk stratification monitoring. The 12-lead electrocardiogram (ECG) is a widely accessible alternative, but its utilization in identifying LV scar has been limited by the complexity and heterogeneity of HCM, even for human experts. To address this challenge, we propose XplainScar, an innovative and explainable machine learning framework that identifies LV scar from 12-lead ECGs. XplainScar employs three key strategies:(1) extracting simple yet comprehensive ECG features to enable explainable predictions of scar in different LV regions \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('337','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_337\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Myocardial scar in the left ventricular contributes significantly to sudden cardiac death in hypertrophic cardiomyopathy (HCM). Although late gadolinium-contrast enhanced magnetic resonance imaging is commonly used for detecting HCM scar, its high cost, limited availability, and susceptibility to artifacts from implanted devices make it unsuitable for ongoing scar progression and risk stratification monitoring. The 12-lead electrocardiogram (ECG) is a widely accessible alternative, but its utilization in identifying LV scar has been limited by the complexity and heterogeneity of HCM, even for human experts. To address this challenge, we propose XplainScar, an innovative and explainable machine learning framework that identifies LV scar from 12-lead ECGs. XplainScar employs three key strategies:(1) extracting simple yet comprehensive ECG features to enable explainable predictions of scar in different LV regions \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('337','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Greenspan, Hayit;  Taylor, Russell;  Madabhushi, Anant;  Deike-Hofmann, Katerina;  Syeda-Mahmood, Tanveer;  Radbruch, Alexander;  Pinetz, Thomas;  Mousavi, Parvin;  Duncan, James;  Kobler, Erich;  Salcudean, Septimiu;  Effland, Alexander;  Haase, Robert<\/p><p class=\"tp_pub_title\">Faithful Synthesis of Low-Dose Contrast-Enhanced Brain MRI Scans Using Noise-Preserving Conditional GANs <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_number\">no. DZNE-2023-01039, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_336\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('336','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_336\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('336','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_336\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{greenspan2023a,<br \/>\r\ntitle = {Faithful Synthesis of Low-Dose Contrast-Enhanced Brain MRI Scans Using Noise-Preserving Conditional GANs},<br \/>\r\nauthor = {Hayit Greenspan and Russell Taylor and Anant Madabhushi and Katerina Deike-Hofmann and Tanveer Syeda-Mahmood and Alexander Radbruch and Thomas Pinetz and Parvin Mousavi and James Duncan and Erich Kobler and Septimiu Salcudean and Alexander Effland and Robert Haase},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-01-01},<br \/>\r\nnumber = {DZNE-2023-01039},<br \/>\r\npublisher = {Clinical Neuroimaging},<br \/>\r\nabstract = {Today Gadolinium-based contrast agents (GBCA) are indispensable in Magnetic Resonance Imaging (MRI) for diagnosing various diseases. However, GBCAs are expensive and may accumulate in patients with potential side effects, thus dose-reduction is recommended. Still, it is unclear to which extent the GBCA dose can be reduced while preserving the diagnostic value\u2013especially in pathological regions. To address this issue, we collected brain MRI scans at numerous non-standard GBCA dosages and developed a conditional GAN model for synthesizing corresponding images at fractional dose levels. Along with the adversarial loss, we advocate a novel content loss function based on the Wasserstein distance of locally paired patch statistics for the faithful preservation of noise. Our numerical experiments show that conditional GANs are suitable for generating images at different GBCA dose levels and can be \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('336','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_336\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Today Gadolinium-based contrast agents (GBCA) are indispensable in Magnetic Resonance Imaging (MRI) for diagnosing various diseases. However, GBCAs are expensive and may accumulate in patients with potential side effects, thus dose-reduction is recommended. Still, it is unclear to which extent the GBCA dose can be reduced while preserving the diagnostic value\u2013especially in pathological regions. To address this issue, we collected brain MRI scans at numerous non-standard GBCA dosages and developed a conditional GAN model for synthesizing corresponding images at fractional dose levels. Along with the adversarial loss, we advocate a novel content loss function based on the Wasserstein distance of locally paired patch statistics for the faithful preservation of noise. Our numerical experiments show that conditional GANs are suitable for generating images at different GBCA dose levels and can be \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('336','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Morton, David;  Connolly, Laura;  Groves, Leah;  Sunderland, Kyle;  Ungi, Tamas;  Jamzad, Amoon;  Kaufmann, Martin;  Ren, Kevin;  Rudan, John F;  Fichtinger, Gabor;  Mousavi, Parvin<\/p><p class=\"tp_pub_title\">Development of a Research Testbed for Intraoperative Optical Spectroscopy Tumor Margin Assessment <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">Acta Polytechnica Hungarica, <\/span><span class=\"tp_pub_additional_volume\">vol. 20, <\/span><span class=\"tp_pub_additional_number\">no. 8, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_335\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('335','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_335\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('335','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_335\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{morton2023b,<br \/>\r\ntitle = {Development of a Research Testbed for Intraoperative Optical Spectroscopy Tumor Margin Assessment},<br \/>\r\nauthor = {David Morton and Laura Connolly and Leah Groves and Kyle Sunderland and Tamas Ungi and Amoon Jamzad and Martin Kaufmann and Kevin Ren and John F Rudan and Gabor Fichtinger and Parvin Mousavi},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-01-01},<br \/>\r\njournal = {Acta Polytechnica Hungarica},<br \/>\r\nvolume = {20},<br \/>\r\nnumber = {8},<br \/>\r\nabstract = {Surgical intervention is a primary treatment option for early-stage cancers. However, the difficulty of intraoperative tumor margin assessment contributes to a high rate of incomplete tumor resection, necessitating revision surgery. This work aims to develop and evaluate a prototype of a tracked tissue sensing research testbed for navigated tumor margin assessment. Our testbed employs diffuse reflection broadband optical spectroscopy for tissue characterization and electromagnetic tracking for navigation. Spectroscopy data and a trained classifier are used to predict tissue types. Navigation allows these predictions to be superimposed on the scanned tissue, creating a spatial classification map. We evaluate the real-time operation of our testbed using an ex vivo tissue phantom. Furthermore, we use the testbed to interrogate ex vivo human kidney tissue and establish a modeling pipeline to classify cancerous and non-neoplastic tissue. The testbed recorded latencies of 125\u00b111 ms and 167\u00b126 ms for navigation and classification respectively. The testbed achieved a Dice similarity coefficient of 93%, and an accuracy of 94% for the spatial classification. These results demonstrated the capabilities of our testbed for the real-time interrogation of an arbitrary tissue volume. Our modeling pipeline attained a balanced accuracy of 91%\u00b14% on the classification of cancerous and non-neoplastic human kidney tissue. Our tracked tissue sensing research testbed prototype shows potential for facilitating the development and evaluation of intraoperative tumor margin assessment technologies across tissue types. The capacity to assess tumor margin status \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('335','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_335\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Surgical intervention is a primary treatment option for early-stage cancers. However, the difficulty of intraoperative tumor margin assessment contributes to a high rate of incomplete tumor resection, necessitating revision surgery. This work aims to develop and evaluate a prototype of a tracked tissue sensing research testbed for navigated tumor margin assessment. Our testbed employs diffuse reflection broadband optical spectroscopy for tissue characterization and electromagnetic tracking for navigation. Spectroscopy data and a trained classifier are used to predict tissue types. Navigation allows these predictions to be superimposed on the scanned tissue, creating a spatial classification map. We evaluate the real-time operation of our testbed using an ex vivo tissue phantom. Furthermore, we use the testbed to interrogate ex vivo human kidney tissue and establish a modeling pipeline to classify cancerous and non-neoplastic tissue. The testbed recorded latencies of 125\u00b111 ms and 167\u00b126 ms for navigation and classification respectively. The testbed achieved a Dice similarity coefficient of 93%, and an accuracy of 94% for the spatial classification. These results demonstrated the capabilities of our testbed for the real-time interrogation of an arbitrary tissue volume. Our modeling pipeline attained a balanced accuracy of 91%\u00b14% on the classification of cancerous and non-neoplastic human kidney tissue. Our tracked tissue sensing research testbed prototype shows potential for facilitating the development and evaluation of intraoperative tumor margin assessment technologies across tissue types. The capacity to assess tumor margin status \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('335','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Chen, Brian;  Maslove, David M;  Curran, Jeffrey D;  Hamilton, Alexander;  Laird, Philip R;  Mousavi, Parvin;  Sibley, Stephanie<\/p><p class=\"tp_pub_title\">A deep learning model for the classification of atrial fibrillation in critically ill patients <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">Intensive care medicine experimental, <\/span><span class=\"tp_pub_additional_volume\">vol. 11, <\/span><span class=\"tp_pub_additional_number\">no. 1, <\/span><span class=\"tp_pub_additional_pages\">pp. 2, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_334\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('334','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_334\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('334','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_334\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{chen2023,<br \/>\r\ntitle = {A deep learning model for the classification of atrial fibrillation in critically ill patients},<br \/>\r\nauthor = {Brian Chen and David M Maslove and Jeffrey D Curran and Alexander Hamilton and Philip R Laird and Parvin Mousavi and Stephanie Sibley},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-01-01},<br \/>\r\njournal = {Intensive care medicine experimental},<br \/>\r\nvolume = {11},<br \/>\r\nnumber = {1},<br \/>\r\npages = {2},<br \/>\r\npublisher = {Springer International Publishing},<br \/>\r\nabstract = {Background <br \/>\r\nAtrial fibrillation (AF) is the most common cardiac arrhythmia in the intensive care unit and is associated with increased morbidity and mortality. New-onset atrial fibrillation (NOAF) is often initially paroxysmal and fleeting, making it difficult to diagnose, and therefore difficult to understand the true burden of disease. Automated algorithms to detect AF in the ICU have been advocated as a means to better quantify its true burden. <br \/>\r\nResults <br \/>\r\nWe used a publicly available 12-lead ECG dataset to train a deep learning model for the classification of AF. We then conducted an external independent validation of the model using continuous telemetry data from 984 critically ill patients collected in our institutional database. Performance metrics were stratified by signal quality, classified as either clean or noisy. The deep learning model was able to classify AF with an overall sensitivity of 84%, specificity of 89%, positive \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('334','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_334\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Background <br \/>\r\nAtrial fibrillation (AF) is the most common cardiac arrhythmia in the intensive care unit and is associated with increased morbidity and mortality. New-onset atrial fibrillation (NOAF) is often initially paroxysmal and fleeting, making it difficult to diagnose, and therefore difficult to understand the true burden of disease. Automated algorithms to detect AF in the ICU have been advocated as a means to better quantify its true burden. <br \/>\r\nResults <br \/>\r\nWe used a publicly available 12-lead ECG dataset to train a deep learning model for the classification of AF. We then conducted an external independent validation of the model using continuous telemetry data from 984 critically ill patients collected in our institutional database. Performance metrics were stratified by signal quality, classified as either clean or noisy. The deep learning model was able to classify AF with an overall sensitivity of 84%, specificity of 89%, positive \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('334','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Kaufmann, Martin;  Iaboni, Natasha;  Jamzad, Amoon;  Hurlbut, David;  Ren, Kevin Yi Mi;  Rudan, John F;  Mousavi, Parvin;  Fichtinger, Gabor;  Varma, Sonal;  Caycedo-Marulanda, Antonio;  Nicol, Christopher JB<\/p><p class=\"tp_pub_title\">Metabolically active zones involving fatty acid elongation delineated by DESI-MSI correlate with pathological and prognostic features of colorectal cancer <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">Metabolites, <\/span><span class=\"tp_pub_additional_volume\">vol. 13, <\/span><span class=\"tp_pub_additional_number\">no. 4, <\/span><span class=\"tp_pub_additional_pages\">pp. 508, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_333\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('333','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_333\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('333','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_333\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{kaufmann2023,<br \/>\r\ntitle = {Metabolically active zones involving fatty acid elongation delineated by DESI-MSI correlate with pathological and prognostic features of colorectal cancer},<br \/>\r\nauthor = {Martin Kaufmann and Natasha Iaboni and Amoon Jamzad and David Hurlbut and Kevin Yi Mi Ren and John F Rudan and Parvin Mousavi and Gabor Fichtinger and Sonal Varma and Antonio Caycedo-Marulanda and Christopher JB Nicol},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-01-01},<br \/>\r\njournal = {Metabolites},<br \/>\r\nvolume = {13},<br \/>\r\nnumber = {4},<br \/>\r\npages = {508},<br \/>\r\npublisher = {MDPI},<br \/>\r\nabstract = {Colorectal cancer (CRC) is the second leading cause of cancer deaths. Despite recent advances, five-year survival rates remain largely unchanged. Desorption electrospray ionization mass spectrometry imaging (DESI) is an emerging nondestructive metabolomics-based method that retains the spatial orientation of small-molecule profiles on tissue sections, which may be validated by \u2018gold standard\u2019 histopathology. In this study, CRC samples were analyzed by DESI from 10 patients undergoing surgery at Kingston Health Sciences Center. The spatial correlation of the mass spectral profiles was compared with histopathological annotations and prognostic biomarkers. Fresh frozen sections of representative colorectal cross sections and simulated endoscopic biopsy samples containing tumour and non-neoplastic mucosa for each patient were generated and analyzed by DESI in a blinded fashion. Sections were then hematoxylin and eosin (H and E) stained, annotated by two independent pathologists, and analyzed. Using PCA\/LDA-based models, DESI profiles of the cross sections and biopsies achieved 97% and 75% accuracies in identifying the presence of adenocarcinoma, using leave-one-patient-out cross validation. Among the m\/z ratios exhibiting the greatest differential abundance in adenocarcinoma were a series of eight long-chain or very-long-chain fatty acids, consistent with molecular and targeted metabolomics indicators of de novo lipogenesis in CRC tissue. Sample stratification based on the presence of lympovascular invasion (LVI), a poor CRC prognostic indicator, revealed the abundance of oxidized phospholipids, suggestive \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('333','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_333\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Colorectal cancer (CRC) is the second leading cause of cancer deaths. Despite recent advances, five-year survival rates remain largely unchanged. Desorption electrospray ionization mass spectrometry imaging (DESI) is an emerging nondestructive metabolomics-based method that retains the spatial orientation of small-molecule profiles on tissue sections, which may be validated by \u2018gold standard\u2019 histopathology. In this study, CRC samples were analyzed by DESI from 10 patients undergoing surgery at Kingston Health Sciences Center. The spatial correlation of the mass spectral profiles was compared with histopathological annotations and prognostic biomarkers. Fresh frozen sections of representative colorectal cross sections and simulated endoscopic biopsy samples containing tumour and non-neoplastic mucosa for each patient were generated and analyzed by DESI in a blinded fashion. Sections were then hematoxylin and eosin (H and E) stained, annotated by two independent pathologists, and analyzed. Using PCA\/LDA-based models, DESI profiles of the cross sections and biopsies achieved 97% and 75% accuracies in identifying the presence of adenocarcinoma, using leave-one-patient-out cross validation. Among the m\/z ratios exhibiting the greatest differential abundance in adenocarcinoma were a series of eight long-chain or very-long-chain fatty acids, consistent with molecular and targeted metabolomics indicators of de novo lipogenesis in CRC tissue. Sample stratification based on the presence of lympovascular invasion (LVI), a poor CRC prognostic indicator, revealed the abundance of oxidized phospholipids, suggestive \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('333','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Kitner, Nicole;  Rodgers, Jessica R;  Ungi, Tamas;  Korzeniowski, Martin;  Olding, Timothy;  Mousavi, Parvin;  Fichtinger, Gabor<\/p><p class=\"tp_pub_title\">Multi-catheter modelling in reconstructed 3D transrectal ultrasound images from prostate brachytherapy <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_pages\">pp. 126-135, <\/span><span class=\"tp_pub_additional_publisher\">SPIE, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_332\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('332','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_332\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('332','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_332\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{kitner2023,<br \/>\r\ntitle = {Multi-catheter modelling in reconstructed 3D transrectal ultrasound images from prostate brachytherapy},<br \/>\r\nauthor = {Nicole Kitner and Jessica R Rodgers and Tamas Ungi and Martin Korzeniowski and Timothy Olding and Parvin Mousavi and Gabor Fichtinger},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-01-01},<br \/>\r\nvolume = {12466},<br \/>\r\npages = {126-135},<br \/>\r\npublisher = {SPIE},<br \/>\r\nabstract = {High-dose-rate brachytherapy is an accepted standard-of-care treatment for prostate cancer. In this procedure, catheters are inserted using three-dimensional (3D) transrectal ultrasound image-guidance. Their positions are manually segmented for treatment planning and delivery. The transverse ultrasound sweep, which is subject to tip and depth error for catheter localization, is a commonly used ultrasound imaging option available for image acquisition. We propose a two-step pipeline that uses a deep-learning network and curve fitting to automatically localize and model catheters in transversely reconstructed 3D ultrasound images. In the first step, a 3D U-Net was trained to automatically segment all catheters in a 3D ultrasound image. Following this step, curve fitting was implemented to detect the shapes of individual catheters using polynomial fitting. Of the 343 catheters (from 20 patients) in the testing data, the \u2026},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('332','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_332\" style=\"display:none;\"><div class=\"tp_abstract_entry\">High-dose-rate brachytherapy is an accepted standard-of-care treatment for prostate cancer. In this procedure, catheters are inserted using three-dimensional (3D) transrectal ultrasound image-guidance. Their positions are manually segmented for treatment planning and delivery. The transverse ultrasound sweep, which is subject to tip and depth error for catheter localization, is a commonly used ultrasound imaging option available for image acquisition. We propose a two-step pipeline that uses a deep-learning network and curve fitting to automatically localize and model catheters in transversely reconstructed 3D ultrasound images. In the first step, a 3D U-Net was trained to automatically segment all catheters in a 3D ultrasound image. Following this step, curve fitting was implemented to detect the shapes of individual catheters using polynomial fitting. Of the 343 catheters (from 20 patients) in the testing data, the \u2026<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('332','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><\/div><div class=\"tablenav\"><div class=\"tablenav-pages\"><span class=\"displaying-num\">337 entries<\/span> <a class=\"page-numbers button disabled\">&laquo;<\/a> <a class=\"page-numbers button disabled\">&lsaquo;<\/a> 1 of 17 <a href=\"https:\/\/labs.cs.queensu.ca\/medilab\/publications\/?limit=2&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=\" title=\"next page\" class=\"page-numbers button\">&rsaquo;<\/a> <a href=\"https:\/\/labs.cs.queensu.ca\/medilab\/publications\/?limit=17&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=\" title=\"last page\" class=\"page-numbers button\">&raquo;<\/a> <\/div><\/div><\/div>\n","protected":false},"excerpt":{"rendered":"","protected":false},"author":2,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"_acf_changed":false,"_uag_custom_page_level_css":"","site-sidebar-layout":"default","site-content-layout":"","ast-site-content-layout":"default","site-content-style":"default","site-sidebar-style":"default","ast-global-header-display":"","ast-banner-title-visibility":"","ast-main-header-display":"","ast-hfb-above-header-display":"","ast-hfb-below-header-display":"","ast-hfb-mobile-header-display":"","site-post-title":"","ast-breadcrumbs-content":"","ast-featured-img":"","footer-sml-layout":"","ast-disable-related-posts":"","theme-transparent-header-meta":"","adv-header-id-meta":"","stick-header-meta":"","header-above-stick-meta":"","header-main-stick-meta":"","header-below-stick-meta":"","astra-migrate-meta-layouts":"default","ast-page-background-enabled":"default","ast-page-background-meta":{"desktop":{"background-color":"var(--ast-global-color-4)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"ast-content-background-meta":{"desktop":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"footnotes":""},"class_list":["post-54","page","type-page","status-publish","hentry"],"acf":[],"spectra_custom_meta":{"_edit_last":["37"],"siteorigin_page_settings":["a:6:{s:6:\"layout\";s:7:\"default\";s:7:\"overlap\";s:8:\"disabled\";s:13:\"header_margin\";b:1;s:10:\"page_title\";b:1;s:13:\"footer_margin\";b:1;s:14:\"footer_widgets\";b:1;}"],"panels_data":["a:3:{s:7:\"widgets\";a:1:{i:0;a:7:{s:5:\"title\";s:0:\"\";s:4:\"text\";s:42:\"[tpsearch]\n\n[tplist author=\"25,57,231,40\"]\";s:6:\"filter\";b:1;s:6:\"visual\";b:1;s:22:\"so_sidebar_emulator_id\";s:12:\"text-5410000\";s:11:\"option_name\";s:11:\"widget_text\";s:11:\"panels_info\";a:7:{s:5:\"class\";s:14:\"WP_Widget_Text\";s:3:\"raw\";b:0;s:4:\"grid\";i:0;s:4:\"cell\";i:0;s:2:\"id\";i:0;s:9:\"widget_id\";s:36:\"02ffda67-9775-45b5-84b9-44ef57ee634d\";s:5:\"style\";a:2:{s:27:\"background_image_attachment\";b:0;s:18:\"background_display\";s:4:\"tile\";}}}}s:5:\"grids\";a:1:{i:0;a:4:{s:5:\"cells\";i:1;s:5:\"style\";a:3:{s:27:\"background_image_attachment\";b:0;s:18:\"background_display\";s:4:\"tile\";s:14:\"cell_alignment\";s:10:\"flex-start\";}s:5:\"ratio\";i:1;s:15:\"ratio_direction\";s:5:\"right\";}}s:10:\"grid_cells\";a:1:{i:0;a:4:{s:4:\"grid\";i:0;s:5:\"index\";i:0;s:6:\"weight\";i:1;s:5:\"style\";a:0:{}}}}"],"um_content_restriction":["a:8:{s:26:\"_um_custom_access_settings\";s:1:\"0\";s:14:\"_um_accessible\";s:1:\"0\";s:19:\"_um_noaccess_action\";s:1:\"0\";s:30:\"_um_restrict_by_custom_message\";s:1:\"0\";s:27:\"_um_restrict_custom_message\";s:0:\"\";s:19:\"_um_access_redirect\";s:1:\"0\";s:23:\"_um_access_redirect_url\";s:0:\"\";s:28:\"_um_access_hide_from_queries\";s:1:\"0\";}"],"_edit_lock":["1761332981:50"],"_uag_css_file_name":["uag-css-54.css"],"_uag_page_assets":["a:9:{s:3:\"css\";s:263:\".uag-blocks-common-selector{z-index:var(--z-index-desktop) !important}@media (max-width: 976px){.uag-blocks-common-selector{z-index:var(--z-index-tablet) !important}}@media (max-width: 767px){.uag-blocks-common-selector{z-index:var(--z-index-mobile) !important}}\n\";s:2:\"js\";s:0:\"\";s:18:\"current_block_list\";a:8:{i:0;s:14:\"core\/shortcode\";i:1;s:11:\"core\/search\";i:2;s:10:\"core\/group\";i:3;s:12:\"core\/heading\";i:4;s:17:\"core\/latest-posts\";i:5;s:20:\"core\/latest-comments\";i:6;s:13:\"core\/archives\";i:7;s:15:\"core\/categories\";}s:8:\"uag_flag\";b:0;s:11:\"uag_version\";s:10:\"1771007182\";s:6:\"gfonts\";a:1:{s:7:\"Default\";a:2:{s:10:\"fontfamily\";s:7:\"Default\";s:12:\"fontvariants\";a:0:{}}}s:10:\"gfonts_url\";s:0:\"\";s:12:\"gfonts_files\";a:0:{}s:14:\"uag_faq_layout\";b:0;}"]},"uagb_featured_image_src":{"full":false,"thumbnail":false,"medium":false,"medium_large":false,"large":false,"1536x1536":false,"2048x2048":false},"uagb_author_info":{"display_name":"Doug Martin","author_link":"https:\/\/labs.cs.queensu.ca\/medilab\/author\/dm70\/"},"uagb_comment_info":0,"uagb_excerpt":null,"_links":{"self":[{"href":"https:\/\/labs.cs.queensu.ca\/medilab\/wp-json\/wp\/v2\/pages\/54","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/labs.cs.queensu.ca\/medilab\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/labs.cs.queensu.ca\/medilab\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/labs.cs.queensu.ca\/medilab\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/labs.cs.queensu.ca\/medilab\/wp-json\/wp\/v2\/comments?post=54"}],"version-history":[{"count":3,"href":"https:\/\/labs.cs.queensu.ca\/medilab\/wp-json\/wp\/v2\/pages\/54\/revisions"}],"predecessor-version":[{"id":2198,"href":"https:\/\/labs.cs.queensu.ca\/medilab\/wp-json\/wp\/v2\/pages\/54\/revisions\/2198"}],"wp:attachment":[{"href":"https:\/\/labs.cs.queensu.ca\/medilab\/wp-json\/wp\/v2\/media?parent=54"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}