From d9bc8127db09f1dd497f7922da3a4216a4a1d97a Mon Sep 17 00:00:00 2001 From: Kyle Date: Thu, 12 Feb 2026 12:58:09 -0500 Subject: [PATCH 1/2] Support NeuroConv 0.6.3 Pin neuroconv to 0.6.3 in all environment files. v0.6.3 is identical to v0.6.2 (re-release for automated upload testing). --- environments/environment-Linux.yml | 4 ++-- environments/environment-MAC-apple-silicon.yml | 4 ++-- environments/environment-MAC-intel.yml | 4 ++-- environments/environment-Windows.yml | 4 ++-- 4 files changed, 8 insertions(+), 8 deletions(-) diff --git a/environments/environment-Linux.yml b/environments/environment-Linux.yml index 297fa56d2..4def6433e 100644 --- a/environments/environment-Linux.yml +++ b/environments/environment-Linux.yml @@ -15,8 +15,8 @@ dependencies: - flask-cors == 4.0.0 - flask_restx == 1.1.0 - werkzeug < 3.0 # werkzeug 3.0 deprecates features used by flask 2.3.2. Remove this when updating flask. - - neuroconv[dandi,compressors] == 0.6.2 - - dandi < 0.74.0 # 0.74.0 renamed dandi-staging to dandi-sandbox, breaking neuroconv 0.6.2 + - neuroconv[dandi,compressors] == 0.6.3 + - dandi < 0.74.0 # 0.74.0 renamed dandi-staging to dandi-sandbox, breaking neuroconv 0.6.3 - spikeinterface >= 0.101.0 # Previously included via neuroconv[ecephys]; needed for tutorial data generation - pandas < 3.0 # pandas 3.0 uses Arrow backend by default, returning read-only arrays that break spikeinterface Phy extractor - neo == 0.14.1 # 0.14.2 is not compatible with neuroconv < 0.7.5 diff --git a/environments/environment-MAC-apple-silicon.yml b/environments/environment-MAC-apple-silicon.yml index ef3983605..9c04ab1f9 100644 --- a/environments/environment-MAC-apple-silicon.yml +++ b/environments/environment-MAC-apple-silicon.yml @@ -23,8 +23,8 @@ dependencies: - werkzeug < 3.0 # werkzeug 3.0 deprecates features used by flask 2.3.2. Remove this when updating flask. # NOTE: the NeuroConv wheel on PyPI includes sonpy which is not compatible with arm64, so build and install # NeuroConv from GitHub, which will remove the sonpy dependency when building from Mac arm64 - - neuroconv[dandi,compressors] == 0.6.2 - - dandi < 0.74.0 # 0.74.0 renamed dandi-staging to dandi-sandbox, breaking neuroconv 0.6.2 + - neuroconv[dandi,compressors] == 0.6.3 + - dandi < 0.74.0 # 0.74.0 renamed dandi-staging to dandi-sandbox, breaking neuroconv 0.6.3 - spikeinterface >= 0.101.0 # Previously included via neuroconv[ecephys]; needed for tutorial data generation - pandas < 3.0 # pandas 3.0 uses Arrow backend by default, returning read-only arrays that break spikeinterface Phy extractor - neo == 0.14.1 # 0.14.2 is not compatible with neuroconv < 0.7.5 diff --git a/environments/environment-MAC-intel.yml b/environments/environment-MAC-intel.yml index 4d33653bb..852651917 100644 --- a/environments/environment-MAC-intel.yml +++ b/environments/environment-MAC-intel.yml @@ -18,8 +18,8 @@ dependencies: - flask-cors == 4.0.0 - flask_restx == 1.1.0 - werkzeug < 3.0 # werkzeug 3.0 deprecates features used by flask 2.3.2. Remove this when updating flask. - - neuroconv[dandi,compressors] == 0.6.2 - - dandi < 0.74.0 # 0.74.0 renamed dandi-staging to dandi-sandbox, breaking neuroconv 0.6.2 + - neuroconv[dandi,compressors] == 0.6.3 + - dandi < 0.74.0 # 0.74.0 renamed dandi-staging to dandi-sandbox, breaking neuroconv 0.6.3 - spikeinterface >= 0.101.0 # Previously included via neuroconv[ecephys]; needed for tutorial data generation - pandas < 3.0 # pandas 3.0 uses Arrow backend by default, returning read-only arrays that break spikeinterface Phy extractor - neo == 0.14.1 # 0.14.2 is not compatible with neuroconv < 0.7.5 diff --git a/environments/environment-Windows.yml b/environments/environment-Windows.yml index 38fd821f4..89c317bdb 100644 --- a/environments/environment-Windows.yml +++ b/environments/environment-Windows.yml @@ -18,8 +18,8 @@ dependencies: - flask-cors === 3.0.10 - flask_restx == 1.1.0 - werkzeug < 3.0 # werkzeug 3.0 deprecates features used by flask 2.3.2. Remove this when updating flask. - - neuroconv[dandi,compressors] == 0.6.2 - - dandi < 0.74.0 # 0.74.0 renamed dandi-staging to dandi-sandbox, breaking neuroconv 0.6.2 + - neuroconv[dandi,compressors] == 0.6.3 + - dandi < 0.74.0 # 0.74.0 renamed dandi-staging to dandi-sandbox, breaking neuroconv 0.6.3 - spikeinterface >= 0.101.0 # Previously included via neuroconv[ecephys]; needed for tutorial data generation - pandas < 3.0 # pandas 3.0 uses Arrow backend by default, returning read-only arrays that break spikeinterface Phy extractor - neo == 0.14.1 # 0.14.2 is not compatible with neuroconv < 0.7.5 From 4d5b0355abedc482f77d251215b44cda5c9199a8 Mon Sep 17 00:00:00 2001 From: Kyle Date: Thu, 12 Feb 2026 13:16:10 -0500 Subject: [PATCH 2/2] Also set DANDI_STAGING_API_KEY for dandi<0.74 compatibility dandi 0.73.x uses instance name 'dandi-staging' which maps to env var DANDI_STAGING_API_KEY. dandi 0.74+ uses 'dandi-sandbox' which maps to DANDI_SANDBOX_API_KEY. Set both for compatibility. --- src/pyflask/manageNeuroconv/manage_neuroconv.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/src/pyflask/manageNeuroconv/manage_neuroconv.py b/src/pyflask/manageNeuroconv/manage_neuroconv.py index c968ee3a8..1e6771bbc 100644 --- a/src/pyflask/manageNeuroconv/manage_neuroconv.py +++ b/src/pyflask/manageNeuroconv/manage_neuroconv.py @@ -1370,6 +1370,7 @@ def upload_folder_to_dandi( os.environ["DANDI_API_KEY"] = api_key if sandbox: os.environ["DANDI_SANDBOX_API_KEY"] = api_key + os.environ["DANDI_STAGING_API_KEY"] = api_key if ignore_cache: os.environ["DANDI_CACHE"] = "ignore" @@ -1403,6 +1404,7 @@ def upload_project_to_dandi( os.environ["DANDI_API_KEY"] = api_key if sandbox: os.environ["DANDI_SANDBOX_API_KEY"] = api_key + os.environ["DANDI_STAGING_API_KEY"] = api_key if ignore_cache: os.environ["DANDI_CACHE"] = "ignore"