Diff of the two buildlogs: -- --- b1/build.log 2025-01-31 11:35:54.716193113 +0000 +++ b2/build.log 2025-01-31 15:28:34.579019973 +0000 @@ -1,6 +1,6 @@ I: pbuilder: network access will be disabled during build -I: Current time: Thu Mar 5 04:21:55 -12 2026 -I: pbuilder-time-stamp: 1772727715 +I: Current time: Sat Feb 1 01:36:04 +14 2025 +I: pbuilder-time-stamp: 1738323364 I: Building the build Environment I: extracting base tarball [/var/cache/pbuilder/unstable-reproducible-base.tgz] I: copying local configuration @@ -48,52 +48,84 @@ dpkg-source: info: applying 1079789_ignore_arm64_nonconvergence.patch I: Not using root during the build. I: Installing the build-deps -I: user script /srv/workspace/pbuilder/1263307/tmp/hooks/D02_print_environment starting +I: user script /srv/workspace/pbuilder/2468496/tmp/hooks/D01_modify_environment starting +debug: Running on codethink04-arm64. +I: Changing host+domainname to test build reproducibility +I: Adding a custom variable just for the fun of it... +I: Changing /bin/sh to bash +'/bin/sh' -> '/bin/bash' +lrwxrwxrwx 1 root root 9 Jan 31 11:36 /bin/sh -> /bin/bash +I: Setting pbuilder2's login shell to /bin/bash +I: Setting pbuilder2's GECOS to second user,second room,second work-phone,second home-phone,second other +I: user script /srv/workspace/pbuilder/2468496/tmp/hooks/D01_modify_environment finished +I: user script /srv/workspace/pbuilder/2468496/tmp/hooks/D02_print_environment starting I: set - BUILDDIR='/build/reproducible-path' - BUILDUSERGECOS='first user,first room,first work-phone,first home-phone,first other' - BUILDUSERNAME='pbuilder1' - BUILD_ARCH='arm64' - DEBIAN_FRONTEND='noninteractive' + BASH=/bin/sh + BASHOPTS=checkwinsize:cmdhist:complete_fullquote:extquote:force_fignore:globasciiranges:globskipdots:hostcomplete:interactive_comments:patsub_replacement:progcomp:promptvars:sourcepath + BASH_ALIASES=() + BASH_ARGC=() + BASH_ARGV=() + BASH_CMDS=() + BASH_LINENO=([0]="12" [1]="0") + BASH_LOADABLES_PATH=/usr/local/lib/bash:/usr/lib/bash:/opt/local/lib/bash:/usr/pkg/lib/bash:/opt/pkg/lib/bash:. + BASH_SOURCE=([0]="/tmp/hooks/D02_print_environment" [1]="/tmp/hooks/D02_print_environment") + BASH_VERSINFO=([0]="5" [1]="2" [2]="37" [3]="1" [4]="release" [5]="aarch64-unknown-linux-gnu") + BASH_VERSION='5.2.37(1)-release' + BUILDDIR=/build/reproducible-path + BUILDUSERGECOS='second user,second room,second work-phone,second home-phone,second other' + BUILDUSERNAME=pbuilder2 + BUILD_ARCH=arm64 + DEBIAN_FRONTEND=noninteractive DEB_BUILD_OPTIONS='buildinfo=+all reproducible=+all parallel=12 ' - DISTRIBUTION='unstable' - HOME='/root' - HOST_ARCH='arm64' + DIRSTACK=() + DISTRIBUTION=unstable + EUID=0 + FUNCNAME=([0]="Echo" [1]="main") + GROUPS=() + HOME=/root + HOSTNAME=i-capture-the-hostname + HOSTTYPE=aarch64 + HOST_ARCH=arm64 IFS=' ' - INVOCATION_ID='532ad5d1dfaf4e52991cb07374cf74c7' - LANG='C' - LANGUAGE='en_US:en' - LC_ALL='C' - MAIL='/var/mail/root' - OPTIND='1' - PATH='/usr/sbin:/usr/bin:/sbin:/bin:/usr/games' - PBCURRENTCOMMANDLINEOPERATION='build' - PBUILDER_OPERATION='build' - PBUILDER_PKGDATADIR='/usr/share/pbuilder' - PBUILDER_PKGLIBDIR='/usr/lib/pbuilder' - PBUILDER_SYSCONFDIR='/etc' - PPID='1263307' - PS1='# ' - PS2='> ' + INVOCATION_ID=ff8e6f2baca345cb872f1588fc1e0638 + LANG=C + LANGUAGE=nl_BE:nl + LC_ALL=C + MACHTYPE=aarch64-unknown-linux-gnu + MAIL=/var/mail/root + OPTERR=1 + OPTIND=1 + OSTYPE=linux-gnu + PATH=/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/i/capture/the/path + PBCURRENTCOMMANDLINEOPERATION=build + PBUILDER_OPERATION=build + PBUILDER_PKGDATADIR=/usr/share/pbuilder + PBUILDER_PKGLIBDIR=/usr/lib/pbuilder + PBUILDER_SYSCONFDIR=/etc + PIPESTATUS=([0]="0") + POSIXLY_CORRECT=y + PPID=2468496 PS4='+ ' - PWD='/' - SHELL='/bin/bash' - SHLVL='2' - SUDO_COMMAND='/usr/bin/timeout -k 18.1h 18h /usr/bin/ionice -c 3 /usr/bin/nice /usr/sbin/pbuilder --build --configfile /srv/reproducible-results/rbuild-debian/r-b-build.8Ukbhx91/pbuilderrc_fFXw --distribution unstable --hookdir /etc/pbuilder/first-build-hooks --debbuildopts -b --basetgz /var/cache/pbuilder/unstable-reproducible-base.tgz --buildresult /srv/reproducible-results/rbuild-debian/r-b-build.8Ukbhx91/b1 --logfile b1/build.log statsmodels_0.14.4+dfsg-1.dsc' - SUDO_GID='109' - SUDO_UID='104' - SUDO_USER='jenkins' - TERM='unknown' - TZ='/usr/share/zoneinfo/Etc/GMT+12' - USER='root' - _='/usr/bin/systemd-run' - http_proxy='http://192.168.101.4:3128' + PWD=/ + SHELL=/bin/bash + SHELLOPTS=braceexpand:errexit:hashall:interactive-comments:posix + SHLVL=3 + SUDO_COMMAND='/usr/bin/timeout -k 24.1h 24h /usr/bin/ionice -c 3 /usr/bin/nice -n 11 /usr/bin/unshare --uts -- /usr/sbin/pbuilder --build --configfile /srv/reproducible-results/rbuild-debian/r-b-build.8Ukbhx91/pbuilderrc_OWwm --distribution unstable --hookdir /etc/pbuilder/rebuild-hooks --debbuildopts -b --basetgz /var/cache/pbuilder/unstable-reproducible-base.tgz --buildresult /srv/reproducible-results/rbuild-debian/r-b-build.8Ukbhx91/b2 --logfile b2/build.log statsmodels_0.14.4+dfsg-1.dsc' + SUDO_GID=109 + SUDO_UID=104 + SUDO_USER=jenkins + TERM=unknown + TZ=/usr/share/zoneinfo/Etc/GMT-14 + UID=0 + USER=root + _='I: set' + http_proxy=http://192.168.101.4:3128 I: uname -a - Linux codethink03-arm64 6.1.0-30-cloud-arm64 #1 SMP Debian 6.1.124-1 (2025-01-12) aarch64 GNU/Linux + Linux i-capture-the-hostname 6.1.0-30-cloud-arm64 #1 SMP Debian 6.1.124-1 (2025-01-12) aarch64 GNU/Linux I: ls -l /bin - lrwxrwxrwx 1 root root 7 Nov 22 2024 /bin -> usr/bin -I: user script /srv/workspace/pbuilder/1263307/tmp/hooks/D02_print_environment finished + lrwxrwxrwx 1 root root 7 Nov 22 14:40 /bin -> usr/bin +I: user script /srv/workspace/pbuilder/2468496/tmp/hooks/D02_print_environment finished -> Attempting to satisfy build-dependencies -> Creating pbuilder-satisfydepends-dummy package Package: pbuilder-satisfydepends-dummy @@ -617,7 +649,7 @@ Get: 400 http://deb.debian.org/debian unstable/main arm64 r-cran-lmtest arm64 0.9.40-1+b1 [403 kB] Get: 401 http://deb.debian.org/debian unstable/main arm64 r-cran-robustbase arm64 0.99-4-1-1 [3042 kB] Get: 402 http://deb.debian.org/debian unstable/main arm64 r-cran-vcd all 1:1.4-13-1 [1287 kB] -Fetched 301 MB in 3s (111 MB/s) +Fetched 301 MB in 1s (259 MB/s) Preconfiguring packages ... Selecting previously unselected package libtext-charwidth-perl:arm64. (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 19956 files and directories currently installed.) @@ -1904,8 +1936,8 @@ Setting up tzdata (2025a-1) ... Current default time zone: 'Etc/UTC' -Local time is now: Thu Mar 5 16:23:30 UTC 2026. -Universal Time is now: Thu Mar 5 16:23:30 UTC 2026. +Local time is now: Fri Jan 31 11:37:01 UTC 2025. +Universal Time is now: Fri Jan 31 11:37:01 UTC 2025. Run 'dpkg-reconfigure tzdata' if you wish to change it. Setting up unicode-data (15.1.0-1) ... @@ -2273,7 +2305,11 @@ Building tag database... -> Finished parsing the build-deps I: Building the package -I: Running cd /build/reproducible-path/statsmodels-0.14.4+dfsg/ && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games" HOME="/nonexistent/first-build" dpkg-buildpackage -us -uc -b && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games" HOME="/nonexistent/first-build" dpkg-genchanges -S > ../statsmodels_0.14.4+dfsg-1_source.changes +I: user script /srv/workspace/pbuilder/2468496/tmp/hooks/A99_set_merged_usr starting +Not re-configuring usrmerge for unstable +I: user script /srv/workspace/pbuilder/2468496/tmp/hooks/A99_set_merged_usr finished +hostname: Name or service not known +I: Running cd /build/reproducible-path/statsmodels-0.14.4+dfsg/ && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/i/capture/the/path" HOME="/nonexistent/second-build" dpkg-buildpackage -us -uc -b && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/i/capture/the/path" HOME="/nonexistent/second-build" dpkg-genchanges -S > ../statsmodels_0.14.4+dfsg-1_source.changes dpkg-buildpackage: info: source package statsmodels dpkg-buildpackage: info: source version 0.14.4+dfsg-1 dpkg-buildpackage: info: source distribution unstable @@ -6121,14 +6157,14 @@ ../tools/nbgenerate.py --parallel --report-errors --skip-existing --execute-only --execution-blacklist statespace_custom_models Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/robust_models_0.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/robust_models_0.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/predict.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/predict.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/plots_boxplots.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/plots_boxplots.ipynbExecuting /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/gee_score_test_simulation.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/gee_score_test_simulation.ipynbExecuting /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/exponential_smoothing.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/exponential_smoothing.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/ets.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/ets.ipynbExecuting /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/copula.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/copula.ipynbExecuting /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/generic_mle.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/generic_mle.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/plots_boxplots.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/plots_boxplots.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/gee_score_test_simulation.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/gee_score_test_simulation.ipynbExecuting /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/exponential_smoothing.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/exponential_smoothing.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/ets.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/ets.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/regression_diagnostics.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/regression_diagnostics.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/copula.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/copula.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/generic_mle.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/generic_mle.ipynbExecuting /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/regression_diagnostics.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/regression_diagnostics.ipynbExecuting /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/mixed_lm_example.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/mixed_lm_example.ipynbExecuting /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/markov_regression.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/markov_regression.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/mixed_lm_example.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/mixed_lm_example.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/markov_regression.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/markov_regression.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/wls.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/wls.ipynb RUNNING THE L-BFGS-B CODE @@ -6197,9 +6233,6 @@ F = 5.3167544390512402 CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/variance_components.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/variance_components.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/tsa_filters.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/tsa_filters.ipynbFinished /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/robust_models_0.ipynb - RUNNING THE L-BFGS-B CODE * * * @@ -6214,6 +6247,7 @@ At iterate 1 f= 2.58255D+00 |proj g|= 9.99244D-01 At iterate 2 f= 2.49918D+00 |proj g|= 2.90033D-01 +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/variance_components.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/variance_components.ipynb At iterate 3 f= 2.48198D+00 |proj g|= 2.44942D-01 @@ -6272,6 +6306,7 @@ At iterate 29 f= 2.29115D+00 |proj g|= 8.46728D-02 At iterate 30 f= 2.27698D+00 |proj g|= 2.33669D-01 +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/tsa_filters.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/tsa_filters.ipynb At iterate 31 f= 2.27536D+00 |proj g|= 1.28536D-01 @@ -6395,17 +6430,17 @@ CONVERGENCE: REL_REDUCTION_OF_F_<=_FACTR*EPSMCH Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/tsa_dates.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/tsa_dates.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/tsa_arma_1.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/tsa_arma_1.ipynbFinished /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/predict.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/tsa_arma_1.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/tsa_arma_1.ipynbFinished /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/robust_models_0.ipynb +Finished /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/predict.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/tsa_arma_0.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/tsa_arma_0.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/treatment_effect.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/treatment_effect.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/theta-model.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/theta-model.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/stl_decomposition.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/stl_decomposition.ipynb Finished /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/plots_boxplots.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/stl_decomposition.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/stl_decomposition.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/stats_rankcompare.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/stats_rankcompare.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/stats_poisson.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/stats_poisson.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/stationarity_detrending_adf_kpss.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/stationarity_detrending_adf_kpss.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_varmax.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_varmax.ipynb ****************************************************************************** ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/theta-model.ipynb @@ -6464,11 +6499,13 @@ ****************************************************************************** +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_varmax.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_varmax.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_tvpvar_mcmc_cfa.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_tvpvar_mcmc_cfa.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_structural_harvey_jaeger.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_structural_harvey_jaeger.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_seasonal.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_seasonal.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_sarimax_stata.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_sarimax_stata.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_sarimax_pymc3.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_sarimax_pymc3.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_sarimax_internet.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_sarimax_internet.ipynb RUNNING THE L-BFGS-B CODE * * * @@ -6500,7 +6537,75 @@ F = 2.2498167187365845 CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_sarimax_internet.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_sarimax_internet.ipynb + +****************************************************************************** +ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_structural_harvey_jaeger.ipynb +An error occurred while executing the following cell: +------------------ +# Datasets +from pandas_datareader.data import DataReader + +# Get the raw data +start = '1948-01' +end = '2008-01' +us_gnp = DataReader('GNPC96', 'fred', start=start, end=end) +us_gnp_deflator = DataReader('GNPDEF', 'fred', start=start, end=end) +us_monetary_base = DataReader('AMBSL', 'fred', start=start, end=end).resample('QS').mean() +recessions = DataReader('USRECQ', 'fred', start=start, end=end).resample('QS').last().values[:,0] + +# Construct the dataframe +dta = pd.concat(map(np.log, (us_gnp, us_gnp_deflator, us_monetary_base)), axis=1) +dta.columns = ['US GNP','US Prices','US monetary base'] +dta.index.freq = dta.index.inferred_freq +dates = dta.index._mpl_repr() +------------------ + + +--------------------------------------------------------------------------- +ModuleNotFoundError Traceback (most recent call last) +Cell In[2], line 2 + 1 # Datasets +----> 2 from pandas_datareader.data import DataReader + 4 # Get the raw data + 5 start = '1948-01' + +ModuleNotFoundError: No module named 'pandas_datareader' + +An error occurred while executing the following cell: +------------------ +# Datasets +from pandas_datareader.data import DataReader + +# Get the raw data +start = '1948-01' +end = '2008-01' +us_gnp = DataReader('GNPC96', 'fred', start=start, end=end) +us_gnp_deflator = DataReader('GNPDEF', 'fred', start=start, end=end) +us_monetary_base = DataReader('AMBSL', 'fred', start=start, end=end).resample('QS').mean() +recessions = DataReader('USRECQ', 'fred', start=start, end=end).resample('QS').last().values[:,0] + +# Construct the dataframe +dta = pd.concat(map(np.log, (us_gnp, us_gnp_deflator, us_monetary_base)), axis=1) +dta.columns = ['US GNP','US Prices','US monetary base'] +dta.index.freq = dta.index.inferred_freq +dates = dta.index._mpl_repr() +------------------ + + +--------------------------------------------------------------------------- +ModuleNotFoundError Traceback (most recent call last) +Cell In[2], line 2 + 1 # Datasets +----> 2 from pandas_datareader.data import DataReader + 4 # Get the raw data + 5 start = '1948-01' + +ModuleNotFoundError: No module named 'pandas_datareader' + +****************************************************************************** + + +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_sarimax_faq.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_sarimax_faq.ipynb ****************************************************************************** ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_varmax.ipynb @@ -6580,12 +6685,12 @@ 215 ) from e 217 sys.audit("http.client.connect", self, self.host, self.port) -NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused +NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused The above exception was the direct cause of the following exception: ProxyError Traceback (most recent call last) -ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) +ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) The above exception was the direct cause of the following exception: @@ -6619,7 +6724,7 @@ --> 519 raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] 521 log.debug("Incremented Retry for (url='%s'): %r", url, new_retry) -MaxRetryError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/lutkepohl2.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) +MaxRetryError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/lutkepohl2.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) During handling of the above exception, another exception occurred: @@ -6653,63 +6758,32 @@ 56 # avoid leaving sockets open which can trigger a ResourceWarning in some 57 # cases, and look like a memory leak in others. 58 with sessions.Session() as session: ----> 59 return session.request(method=method, url=url, **kwargs) - -File /usr/lib/python3/dist-packages/requests/sessions.py:589, in Session.request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json) - 584 send_kwargs = { - 585 "timeout": timeout, - 586 "allow_redirects": allow_redirects, - 587 } - 588 send_kwargs.update(settings) ---> 589 resp = self.send(prep, **send_kwargs) - 591 return resp - -File /usr/lib/python3/dist-packages/requests/sessions.py:703, in Session.send(self, request, **kwargs) - 700 start = preferred_clock() - 702 # Send the request ---> 703 r = adapter.send(request, **kwargs) - 705 # Total elapsed time of the request (approximately) - 706 elapsed = preferred_clock() - start - -File /usr/lib/python3/dist-packages/requests/adapters.py:694, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) - 691 raise RetryError(e, request=request) - 693 if isinstance(e.reason, _ProxyError): ---> 694 raise ProxyError(e, request=request) - 696 if isinstance(e.reason, _SSLError): - 697 # This branch is for urllib3 v1.22 and later. - 698 raise SSLError(e, request=request) - -ProxyError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/lutkepohl2.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) +****************************************************************************** +ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_sarimax_stata.ipynb An error occurred while executing the following cell: ------------------ -import requests -import shutil - -def download_file(url): - local_filename = url.split('/')[-1] - with requests.get(url, stream=True) as r: - with open(local_filename, 'wb') as f: - shutil.copyfileobj(r.raw, f) - - return local_filename - -filename = download_file("https://www.stata-press.com/data/r12/lutkepohl2.dta") +# Dataset +wpi1 = requests.get('https://www.stata-press.com/data/r12/wpi1.dta').content +data = pd.read_stata(BytesIO(wpi1)) +data.index = data.t +# Set the frequency +data.index.freq="QS-OCT" -dta = pd.read_stata(filename) -dta.index = dta.qtr -dta.index.freq = dta.index.inferred_freq -endog = dta.loc['1960-04-01':'1978-10-01', ['dln_inv', 'dln_inc', 'dln_consump']] +# Fit the model +mod = sm.tsa.statespace.SARIMAX(data['wpi'], trend='c', order=(1,1,1)) +res = mod.fit(disp=False) +print(res.summary()) ------------------ ---------------------------------------------------------------------------- +---> 59--------------------------------------------------------------------------- ConnectionRefusedError Traceback (most recent call last) File /usr/lib/python3/dist-packages/urllib3/connection.py:198, in HTTPConnection._new_conn(self) 197 try: --> 198 sock = connection.create_connection( 199 (self._dns_host, self.port), - 200 self.timeout, + 200 self. timeout, 201 source_address=self.source_address, 202 socket_options=self.socket_options, 203 ) @@ -6719,7 +6793,7 @@ 84 try: ---> 85 raise err 86 finally: - 87 # Break explicitly a reference cycle + 87 # Break explicitly a reference cyclereturn File /usr/lib/python3/dist-packages/urllib3/util/connection.py:73, in create_connection(address, timeout, source_address, socket_options) 72 sock.bind(source_address) @@ -6733,7 +6807,7 @@ NewConnectionError Traceback (most recent call last) File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:773, in HTTPConnectionPool.urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw) 772 try: ---> 773 self._prepare_proxy(conn) +--> 773 self._prepare_proxy(conn ) 774 except (BaseSSLError, OSError, SocketTimeout) as e: File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:1042, in HTTPSConnectionPool._prepare_proxy(self, conn) @@ -6741,7 +6815,7 @@ 1037 scheme=tunnel_scheme, 1038 host=self._tunnel_host, 1039 port=self.port, - 1040 headers=self.proxy_headers, + 1040 headers=self.sessionproxy_headers, 1041 ) -> 1042 conn.connect() @@ -6757,17 +6831,17 @@ 215 ) from e 217 sys.audit("http.client.connect", self, self.host, self.port) -NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused +NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused The above exception was the direct cause of the following exception: ProxyError Traceback (most recent call last) -ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) +ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) The above exception was the direct cause of the following exception: MaxRetryError Traceback (most recent call last) -File /usr/lib/python3/dist-packages/requests/adapters.py:667, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) +File /usr/lib/python3/dist-packages/requests/adapters.py:667., in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) 666 try: --> 667 resp = conn.urlopen( 668 method=request.method, @@ -6775,7 +6849,7 @@ 670 body=request.body, 671 headers=request.headers, 672 redirect=False, - 673 assert_same_host=False, + 673 assert_same_host=requestFalse, 674 preload_content=False, 675 decode_content=False, 676 retries=self.max_retries, @@ -6787,7 +6861,7 @@ File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:841, in HTTPConnectionPool.urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw) 839 new_e = ProtocolError("Connection aborted.", new_e) --> 841 retries = retries.increment( - 842 method, url, error=new_e, _pool=self, _stacktrace=sys.exc_info()[2] + 842 method, (url, error=new_e, _pool=self, _stacktrace=sys.exc_info()[2] 843 ) 844 retries.sleep() @@ -6796,24 +6870,16 @@ --> 519 raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] 521 log.debug("Incremented Retry for (url='%s'): %r", url, new_retry) -MaxRetryError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/lutkepohl2.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) +MaxRetryError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/wpi1.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) During handling of the above exception, another exception occurred: -ProxyError Traceback (most recent call last) -Cell In[3], line 12 - 8 shutil.copyfileobj(r.raw, f) - 10 return local_filename ----> 12 filename = download_file("https://www.stata-press.com/data/r12/lutkepohl2.dta") - 14 dta = pd.read_stata(filename) - 15 dta.index = dta.qtr - -Cell In[3], line 6, in download_file(url) - 4 def download_file(url): - 5 local_filename = url.split('/')[-1] -----> 6 with requests.get(url, stream=True) as r: - 7 with open(local_filename, 'wb') as f: - 8 shutil.copyfileobj(r.raw, f) +methodProxyError Traceback (most recent call last) +Cell In[3], line 2 + 1 # Dataset +----> 2 wpi1 = requests.get('https://www.stata-press.com/data/r12/wpi1.dta').content + 3 data = pd.read_stata(BytesIO(wpi1)) + 4 data.index = data.t File /usr/lib/python3/dist-packages/requests/api.py:73, in get(url, params, **kwargs) 62 def get(url, params=None, **kwargs): @@ -6830,7 +6896,7 @@ 56 # avoid leaving sockets open which can trigger a ResourceWarning in some 57 # cases, and look like a memory leak in others. 58 with sessions.Session() as session: ----> 59 return session.request(method=method, url=url, **kwargs) +---> 59 return session.request(=method=method, url=url, **kwargs) File /usr/lib/python3/dist-packages/requests/sessions.py:589, in Session.request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json) 584 send_kwargs = { @@ -6856,150 +6922,8 @@ 697 # This branch is for urllib3 v1.22 and later. 698 raise SSLError(e, request=request) -ProxyError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/lutkepohl2.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) - -****************************************************************************** - - -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_sarimax_faq.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_sarimax_faq.ipynb - -****************************************************************************** -ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_structural_harvey_jaeger.ipynb -An error occurred while executing the following cell: ------------------- -# Datasets -from pandas_datareader.data import DataReader - -# Get the raw data -start = '1948-01' -end = '2008-01' -us_gnp = DataReader('GNPC96', 'fred', start=start, end=end) -us_gnp_deflator = DataReader('GNPDEF', 'fred', start=start, end=end) -us_monetary_base = DataReader('AMBSL', 'fred', start=start, end=end).resample('QS').mean() -recessions = DataReader('USRECQ', 'fred', start=start, end=end).resample('QS').last().values[:,0] - -# Construct the dataframe -dta = pd.concat(map(np.log, (us_gnp, us_gnp_deflator, us_monetary_base)), axis=1) -dta.columns = ['US GNP','US Prices','US monetary base'] -dta.index.freq = dta.index.inferred_freq -dates = dta.index._mpl_repr() ------------------- - - ---------------------------------------------------------------------------- -ModuleNotFoundError Traceback (most recent call last) -Cell In[2], line 2 - 1 # Datasets -----> 2 from pandas_datareader.data import DataReader - 4 # Get the raw data - 5 start = '1948-01' - -ModuleNotFoundError: No module named 'pandas_datareader' - -An error occurred while executing the following cell: ------------------- -# Datasets -from pandas_datareader.data import DataReader - -# Get the raw data -start = '1948-01' -end = '2008-01' -us_gnp = DataReader('GNPC96', 'fred', start=start, end=end) -us_gnp_deflator = DataReader('GNPDEF', 'fred', start=start, end=end) -us_monetary_base = DataReader('AMBSL', 'fred', start=start, end=end).resample('QS').mean() -recessions = DataReader('USRECQ', 'fred', start=start, end=end).resample('QS').last().values[:,0] - -# Construct the dataframe -dta = pd.concat(map(np.log, (us_gnp, us_gnp_deflator, us_monetary_base)), axis=1) -dta.columns = ['US GNP','US Prices','US monetary base'] -dta.index.freq = dta.index.inferred_freq -dates = dta.index._mpl_repr() ------------------- - +ProxyErrormethod: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/wpi1.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) ---------------------------------------------------------------------------- -ModuleNotFoundError Traceback (most recent call last) -Cell In[2], line 2 - 1 # Datasets -----> 2 from pandas_datareader.data import DataReader - 4 # Get the raw data - 5 start = '1948-01' - -ModuleNotFoundError: No module named 'pandas_datareader' - -****************************************************************************** - - -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_news.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_news.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_local_linear_trend.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_local_linear_trend.ipynb - -****************************************************************************** -ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_sarimax_pymc3.ipynb -An error occurred while executing the following cell: ------------------- -%matplotlib inline -import matplotlib.pyplot as plt -import numpy as np -import pandas as pd -import pymc3 as pm -import statsmodels.api as sm -import theano -import theano.tensor as tt -from pandas.plotting import register_matplotlib_converters -from pandas_datareader.data import DataReader - -plt.style.use("seaborn") -register_matplotlib_converters() ------------------- - - ---------------------------------------------------------------------------- -ModuleNotFoundError Traceback (most recent call last) -Cell In[1], line 5 - 3 import numpy as np - 4 import pandas as pd -----> 5 import pymc3 as pm - 6 import statsmodels.api as sm - 7 import theano - -ModuleNotFoundError: No module named 'pymc3' - -An error occurred while executing the following cell: ------------------- -%matplotlib inline -import matplotlib.pyplot as plt -import numpy as np -import pandas as pd -import pymc3 as pm -import statsmodels.api as sm -import theano -import theano.tensor as tt -from pandas.plotting import register_matplotlib_converters -from pandas_datareader.data import DataReader - -plt.style.use("seaborn") -register_matplotlib_converters() ------------------- - - ---------------------------------------------------------------------------- -ModuleNotFoundError Traceback (most recent call last) -Cell In[1], line 5 - 3 import numpy as np - 4 import pandas as pd -----> 5 import pymc3 as pm - 6 import statsmodels.api as sm - 7 import theano - -ModuleNotFoundError: No module named 'pymc3' - -****************************************************************************** - - -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_forecasting.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_forecasting.ipynb - -****************************************************************************** -ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_sarimax_stata.ipynb An error occurred while executing the following cell: ------------------ # Dataset @@ -7022,7 +6946,7 @@ 197 try: --> 198 sock = connection.create_connection( 199 (self._dns_host, self.port), - 200 self.timeout, + 200 self.timeout,, 201 source_address=self.source_address, 202 socket_options=self.socket_options, 203 ) @@ -7036,7 +6960,7 @@ File /usr/lib/python3/dist-packages/urllib3/util/connection.py:73, in create_connection(address, timeout, source_address, socket_options) 72 sock.bind(source_address) ----> 73 sock.connect(sa) +---> 73 sock.connect (sa) 74 # Break explicitly a reference cycle ConnectionRefusedError: [Errno 111] Connection refused @@ -7070,22 +6994,92 @@ 215 ) from e 217 sys.audit("http.client.connect", self, self.host, self.port) -NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused +NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused The above exception was the direct cause of the following exception: ProxyError Traceback (most recent call last) -ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) +ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) The above exception was the direct cause of the following exception: MaxRetryError Traceback (most recent call last) -File /usr/lib/python3/dist-packages/requests/adapters.py:667, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) - 666 try: +File /usr/lib/python3/dist-packages/requests/adapters.py:667, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies)url=url, **kwargs) + +File /usr/lib/python3/dist-packages/requests/sessions.py:589, in Session.request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json) + 584 send_kwargs = { + 585 "timeout": timeout, + 586 "allow_redirects": allow_redirects, + 587 } + 588 send_kwargs.update(settings) +--> 589 resp = self.send(prep, **send_kwargs) + 591 return resp + +File /usr/lib/python3/dist-packages/requests/sessions.py:703, in Session.send(self, request, **kwargs) + 700 start = preferred_clock() + 702 # Send the request +--> 703 r = adapter.send(request, **kwargs) + 705 # Total elapsed time of the request (approximately) + 706 elapsed = preferred_clock() - start + +File /usr/lib/python3/dist-packages/requests/adapters.py:694, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) + 691 raise RetryError(e, request=request) + 693 if isinstance(e.reason, _ProxyError): +--> 694 raise ProxyError(e, request=request) + 696 if isinstance(e.reason, _SSLError): + 697 # This branch is for urllib3 v1.22 and later. + 698 raise SSLError(e, request=request) + +ProxyError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/lutkepohl2.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) + +An error occurred while executing the following cell: +------------------ +import requests +import shutil + +def download_file(url): + local_filename = url.split('/')[-1] + with requests.get(url, stream=True) as r: + with open(local_filename, 'wb') as f: + shutil.copyfileobj(r.raw, f) + + return local_filename + +filename = download_file("https://www.stata-press.com/data/r12/lutkepohl2.dta") + +dta = pd.read_stata(filename) +dta.index = dta.qtr +dta.index.freq = dta.index.inferred_freq +endog = dta.loc['1960-04-01':'1978-10-01', ['dln_inv', 'dln_inc', 'dln_consump']] +------------------ + + +--------------------------------------------------------------------------- +ConnectionRefusedError Traceback (most recent call last) +File /usr/lib/python3/dist-packages/urllib3/connection.py:198, in HTTPConnection._new_conn(self) + 197 try: +--> 198 sock = connection.create_connection( + 199 (self._dns_host, self.port), + 200 self.timeout, + 201 source_address=self.source_address, + 202 socket_options=self.socket_options, + 203 ) + 204 except socket.gaierror as e: + +File /usr/lib/python3/dist-packages/urllib3/util/connection.py:85, in create_connection(address, timeout, source_address, socket_options) + 84 try: +---> 85 raise err + 86 finally: + 87 # Break explicitly a reference cycle + +File /usr/lib/python3/dist-packages/urllib3/util/connection.py:73, in create_connection(address, timeout, source_address, socket_options) + 72 sock.bind(source_address) +---> 73 +sock 666 try: --> 667 resp = conn.urlopen( 668 method=request.method, 669 url=url, - 670 body=request.body, + 670 body=request..body, 671 headers=request.headers, 672 redirect=False, 673 assert_same_host=False, @@ -7093,14 +7087,14 @@ 675 decode_content=False, 676 retries=self.max_retries, 677 timeout=timeout, - 678 chunked=chunked, + 678 chunked=chunkedconnect, 679 ) 681 except (ProtocolError, OSError) as err: File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:841, in HTTPConnectionPool.urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw) 839 new_e = ProtocolError("Connection aborted.", new_e) --> 841 retries = retries.increment( - 842 method, url, error=new_e, _pool=self, _stacktrace=sys.exc_info()[2] + 842 method, url, error=new_e, _pool=self(, _stacktrace=sys.exc_info()[2] 843 ) 844 retries.sleep() @@ -7109,7 +7103,7 @@ --> 519 raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] 521 log.debug("Incremented Retry for (url='%s'): %r", url, new_retry) -MaxRetryError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/wpi1.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) +MaxRetryErrorsa: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/wpi1.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) During handling of the above exception, another exception occurred: @@ -7121,21 +7115,22 @@ 4 data.index = data.t File /usr/lib/python3/dist-packages/requests/api.py:73, in get(url, params, **kwargs) - 62 def get(url, params=None, **kwargs): + 62 def get(url, params=None), **kwargs): 63 r"""Sends a GET request. 64 65 :param url: URL for the new :class:`Request` object. (...) 70 :rtype: requests.Response 71 """ ----> 73 return request("get", url, params=params, **kwargs) +---> 73 return request("get", url, params=params, * +*kwargs) File /usr/lib/python3/dist-packages/requests/api.py:59, in request(method, url, **kwargs) 55 # By using the 'with' statement we are sure the session is closed, thus we 56 # avoid leaving sockets open which can trigger a ResourceWarning in some 57 # cases, and look like a memory leak in others. 58 with sessions.Session() as session: ----> 59 return session.request(method=method, url=url, **kwargs) +---> 59 return session.request(method=method, url=url, 74 **kwargs) File /usr/lib/python3/dist-packages/requests/sessions.py:589, in Session.request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json) 584 send_kwargs = { @@ -7143,13 +7138,13 @@ 586 "allow_redirects": allow_redirects, 587 } 588 send_kwargs.update(settings) ---> 589 resp = self.send(prep, **send_kwargs) +--> 589 resp = self.send( prep, **send_kwargs) 591 return resp File /usr/lib/python3/dist-packages/requests/sessions.py:703, in Session.send(self, request, **kwargs) 700 start = preferred_clock() 702 # Send the request ---> 703 r = adapter.send(request, **kwargs) +--> 703 r = adapter.send(request, **kwargs)# Break explicitly a reference cycle 705 # Total elapsed time of the request (approximately) 706 elapsed = preferred_clock() - start @@ -7161,48 +7156,14 @@ 697 # This branch is for urllib3 v1.22 and later. 698 raise SSLError(e, request=request) -ProxyError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/wpi1.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) - -An error occurred while executing the following cell: ------------------- -# Dataset -wpi1 = requests.get('https://www.stata-press.com/data/r12/wpi1.dta').content -data = pd.read_stata(BytesIO(wpi1)) -data.index = data.t -# Set the frequency -data.index.freq="QS-OCT" - -# Fit the model -mod = sm.tsa.statespace.SARIMAX(data['wpi'], trend='c', order=(1,1,1)) -res = mod.fit(disp=False) -print(res.summary()) ------------------- - +ProxyError ---------------------------------------------------------------------------- -ConnectionRefusedError Traceback (most recent call last) -File /usr/lib/python3/dist-packages/urllib3/connection.py:198, in HTTPConnection._new_conn(self) - 197 try: ---> 198 sock = connection.create_connection( - 199 (self._dns_host, self.port), - 200 self.timeout, - 201 source_address=self.source_address, - 202 socket_options=self.socket_options, - 203 ) - 204 except socket.gaierror as e: - -File /usr/lib/python3/dist-packages/urllib3/util/connection.py:85, in create_connection(address, timeout, source_address, socket_options) - 84 try: ----> 85 raise err - 86 finally: - 87 # Break explicitly a reference cycle +: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/wpi1.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) -File /usr/lib/python3/dist-packages/urllib3/util/connection.py:73, in create_connection(address, timeout, source_address, socket_options) - 72 sock.bind(source_address) ----> 73 sock.connect(sa) - 74 # Break explicitly a reference cycle +****************************************************************************** -ConnectionRefusedError: [Errno 111] Connection refused +ConnectionRefusedError +: [Errno 111] Connection refused The above exception was the direct cause of the following exception: @@ -7233,12 +7194,12 @@ 215 ) from e 217 sys.audit("http.client.connect", self, self.host, self.port) -NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused +NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused The above exception was the direct cause of the following exception: ProxyError Traceback (most recent call last) -ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) +ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) The above exception was the direct cause of the following exception: @@ -7272,16 +7233,24 @@ --> 519 raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] 521 log.debug("Incremented Retry for (url='%s'): %r", url, new_retry) -MaxRetryError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/wpi1.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) +MaxRetryError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/lutkepohl2.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) During handling of the above exception, another exception occurred: ProxyError Traceback (most recent call last) -Cell In[3], line 2 - 1 # Dataset -----> 2 wpi1 = requests.get('https://www.stata-press.com/data/r12/wpi1.dta').content - 3 data = pd.read_stata(BytesIO(wpi1)) - 4 data.index = data.t +Cell In[3], line 12 + 8 shutil.copyfileobj(r.raw, f) + 10 return local_filename +---> 12 filename = download_file("https://www.stata-press.com/data/r12/lutkepohl2.dta") + 14 dta = pd.read_stata(filename) + 15 dta.index = dta.qtr + +Cell In[3], line 6, in download_file(url) + 4 def download_file(url): + 5 local_filename = url.split('/')[-1] +----> 6 with requests.get(url, stream=True) as r: + 7 with open(local_filename, 'wb') as f: + 8 shutil.copyfileobj(r.raw, f) File /usr/lib/python3/dist-packages/requests/api.py:73, in get(url, params, **kwargs) 62 def get(url, params=None, **kwargs): @@ -7324,11 +7293,78 @@ 697 # This branch is for urllib3 v1.22 and later. 698 raise SSLError(e, request=request) -ProxyError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/wpi1.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) +ProxyError: HTTPSConnectionPool(host='www.stata-press.com', port=443): Max retries exceeded with url: /data/r12/lutkepohl2.dta (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) + +****************************************************************************** + + +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_local_linear_trend.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_local_linear_trend.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_news.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_news.ipynb + +****************************************************************************** +ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_sarimax_pymc3.ipynb +An error occurred while executing the following cell: +------------------ +%matplotlib inline +import matplotlib.pyplot as plt +import numpy as np +import pandas as pd +import pymc3 as pm +import statsmodels.api as sm +import theano +import theano.tensor as tt +from pandas.plotting import register_matplotlib_converters +from pandas_datareader.data import DataReader + +plt.style.use("seaborn") +register_matplotlib_converters() +------------------ + + +--------------------------------------------------------------------------- +ModuleNotFoundError Traceback (most recent call last) +Cell In[1], line 5 + 3 import numpy as np + 4 import pandas as pd +----> 5 import pymc3 as pm + 6 import statsmodels.api as sm + 7 import theano + +ModuleNotFoundError: No module named 'pymc3' + +An error occurred while executing the following cell: +------------------ +%matplotlib inline +import matplotlib.pyplot as plt +import numpy as np +import pandas as pd +import pymc3 as pm +import statsmodels.api as sm +import theano +import theano.tensor as tt +from pandas.plotting import register_matplotlib_converters +from pandas_datareader.data import DataReader + +plt.style.use("seaborn") +register_matplotlib_converters() +------------------ + + +--------------------------------------------------------------------------- +ModuleNotFoundError Traceback (most recent call last) +Cell In[1], line 5 + 3 import numpy as np + 4 import pandas as pd +----> 5 import pymc3 as pm + 6 import statsmodels.api as sm + 7 import theano + +ModuleNotFoundError: No module named 'pymc3' ****************************************************************************** +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_forecasting.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_forecasting.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_fixed_params.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_fixed_params.ipynb RUNNING THE L-BFGS-B CODE @@ -7398,9 +7434,7 @@ At X0 0 variables are exactly at the bounds -At iterate 0 f= 2.32873D+00 |proj g|= 8.23649D-03 - -At iterate 5 f= 2.32864D+00 |proj g|= 1.41994D-03 +At iterate 0 f= 1.41373D+00 |proj g|= 1.06920D-04 * * * @@ -7415,63 +7449,320 @@ * * * N Tit Tnf Tnint Skip Nact Projg F - 3 8 10 1 0 0 5.820D-06 2.329D+00 - F = 2.3286389358138648 + 3 1 4 1 0 0 4.752D-05 1.414D+00 + F = 1.4137311099487528 -CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL +CONVERGENCE: REL_REDUCTION_OF_F_<=_FACTR*EPSMCH RUNNING THE L-BFGS-B CODE * * * Machine precision = 2.220D-16 - N = 3 M = 10 + N = 4 M = 10 This problem is unconstrained. At X0 0 variables are exactly at the bounds -At iterate 0 f= 1.41373D+00 |proj g|= 1.06920D-04 +At iterate 0 f= 1.42683D+00 |proj g|= 2.05943D-01 - * * * +At iterate 5 f= 1.41332D+00 |proj g|= 1.60874D-03 -Tit = total number of iterations -Tnf = total number of function evaluations -Tnint = total number of segments explored during Cauchy searches -Skip = number of BFGS updates skipped -Nact = number of active bounds at final generalized Cauchy point -Projg = norm of the final projected gradient -F = final function value +****************************************************************************** +ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_local_linear_trend.ipynb +An error occurred while executing the following cell: +------------------ +import requests +from io import BytesIO +from zipfile import ZipFile + +# Download the dataset +ck = requests.get('http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip').content +zipped = ZipFile(BytesIO(ck)) +df = pd.read_table( + BytesIO(zipped.read('OxCodeIntroStateSpaceBook/Chapter_2/NorwayFinland.txt')), + skiprows=1, header=None, sep='\s+', engine='python', + names=['date','nf', 'ff'] +) +------------------ - * * * +----- stderr ----- +<>:10: SyntaxWarning: invalid escape sequence '\s' +<>:10: SyntaxWarning: invalid escape sequence '\s' +/tmp/ipykernel_3504114/2613092943.py:10: SyntaxWarning: invalid escape sequence '\s' + skiprows=1, header=None, sep='\s+', engine='python', +----- stderr ----- +/tmp/ipykernel_3504114/2613092943.py:10: SyntaxWarning: invalid escape sequence '\s' + skiprows=1, header=None, sep='\s+', engine='python', +------------------ - N Tit Tnf Tnint Skip Nact Projg F - 3 1 4 1 0 0 4.752D-05 1.414D+00 - F = 1.4137311099487528 +--------------------------------------------------------------------------- +ConnectionRefusedError Traceback (most recent call last) +File /usr/lib/python3/dist-packages/urllib3/connection.py:198, in HTTPConnection._new_conn(self) + 197 try: +--> 198 sock = connection.create_connection( + 199 (self._dns_host, self.port), + 200 self.timeout, + 201 source_address=self.source_address, + 202 socket_options=self.socket_options, + 203 ) + 204 except socket.gaierror as e: -CONVERGENCE: REL_REDUCTION_OF_F_<=_FACTR*EPSMCH -RUNNING THE L-BFGS-B CODE +File /usr/lib/python3/dist-packages/urllib3/util/connection.py:85, in create_connection(address, timeout, source_address, socket_options) + 84 try: +---> 85 raise err + 86 finally: + 87 # Break explicitly a reference cycle - * * * +File /usr/lib/python3/dist-packages/urllib3/util/connection.py:73, in create_connection(address, timeout, source_address, socket_options) + 72 sock.bind(source_address) +---> 73 sock.connect(sa) + 74 # Break explicitly a reference cycle -Machine precision = 2.220D-16 - N = 4 M = 10 - This problem is unconstrained. +ConnectionRefusedError: [Errno 111] Connection refused -At X0 0 variables are exactly at the bounds +The above exception was the direct cause of the following exception: -At iterate 0 f= 1.42683D+00 |proj g|= 2.05943D-01 -RUNNING THE L-BFGS-B CODE +NewConnectionError Traceback (most recent call last) +File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:787, in HTTPConnectionPool.urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw) + 786 # Make the request on the HTTPConnection object +--> 787 response = self._make_request( + 788 conn, + 789 method, + 790 url, + 791 timeout=timeout_obj, + 792 body=body, + 793 headers=headers, + 794 chunked=chunked, + 795 retries=retries, + 796 response_conn=response_conn, + 797 preload_content=preload_content, + 798 decode_content=decode_content, + 799 **response_kw, + 800 ) + 802 # Everything went great! - * * * +File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:493, in HTTPConnectionPool._make_request(self, conn, method, url, body, headers, retries, timeout, chunked, response_conn, preload_content, decode_content, enforce_content_length) + 492 try: +--> 493 conn.request( + 494 method, + 495 url, + 496 body=body, + 497 headers=headers, + 498 chunked=chunked, + 499 preload_content=preload_content, + 500 decode_content=decode_content, + 501 enforce_content_length=enforce_content_length, + 502 ) + 504 # We are swallowing BrokenPipeError (errno.EPIPE) since the server is + 505 # legitimately able to close the connection after sending a valid response. + 506 # With this behaviour, the received response is still readable. -Machine precision = 2.220D-16 - N = 3 M = 10 - This problem is unconstrained. +File /usr/lib/python3/dist-packages/urllib3/connection.py:445, in HTTPConnection.request(self, method, url, body, headers, chunked, preload_content, decode_content, enforce_content_length) + 444 self.putheader(header, value) +--> 445 self.endheaders() + 447 # If we're given a body we start sending that in chunks. -At X0 0 variables are exactly at the bounds +File /usr/lib/python3.13/http/client.py:1331, in HTTPConnection.endheaders(self, message_body, encode_chunked) + 1330 raise CannotSendHeader() +-> 1331 self._send_output(message_body, encode_chunked=encode_chunked) -At iterate 0 f= 2.23132D+00 |proj g|= 1.09171D-02 +File /usr/lib/python3.13/http/client.py:1091, in HTTPConnection._send_output(self, message_body, encode_chunked) + 1090 del self._buffer[:] +-> 1091 self.send(msg) + 1093 if message_body is not None: + 1094 + 1095 # create a consistent interface to message_body -At iterate 5 f= 1.41332D+00 |proj g|= 1.60874D-03 +File /usr/lib/python3.13/http/client.py:1035, in HTTPConnection.send(self, data) + 1034 if self.auto_open: +-> 1035 self.connect() + 1036 else: + +File /usr/lib/python3/dist-packages/urllib3/connection.py:276, in HTTPConnection.connect(self) + 275 def connect(self) -> None: +--> 276 self.sock = self._new_conn() + 277 if self._tunnel_host: + 278 # If we're tunneling it means we're connected to our proxy. + +File /usr/lib/python3/dist-packages/urllib3/connection.py:213, in HTTPConnection._new_conn(self) + 212 except OSError as e: +--> 213 raise NewConnectionError( + 214 self, f"Failed to establish a new connection: {e}" + 215 ) from e + 217 sys.audit("http.client.connect", self, self.host, self.port) + +NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused + +The above exception was the direct cause of the following exception: + +ProxyError Traceback (most recent call last) +ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) + +The above exception was the direct cause of the following exception: + +MaxRetryError Traceback (most recent call last) +File /usr/lib/python3/dist-packages/requests/adapters.py:667, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) + 666 try: +--> 667 resp = conn.urlopen( + 668 method=request.method, + 669 url=url, + 670 body=request.body, + 671 headers=request.headers, + 672 redirect=False, + 673 assert_same_host=False, + 674 preload_content=False, + 675 decode_content=False, + 676 retries=self.max_retries, + 677 timeout=timeout, + 678 chunked=chunked, + 679 ) + 681 except (ProtocolError, OSError) as err: + +File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:841, in HTTPConnectionPool.urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw) + 839 new_e = ProtocolError("Connection aborted.", new_e) +--> 841 retries = retries.increment( + 842 method, url, error=new_e, _pool=self, _stacktrace=sys.exc_info()[2] + 843 ) + 844 retries.sleep() + +File /usr/lib/python3/dist-packages/urllib3/util/retry.py:519, in Retry.increment(self, method, url, response, error, _pool, _stacktrace) + 518 reason = error or ResponseError(cause) +--> 519 raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] + 521 log.debug("Incremented Retry for (url='%s'): %r", url, new_retry) + +MaxRetryError: HTTPConnectionPool(host='127.0.0.1', port=9): Max retries exceeded with url: http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) + +During handling of the above exception, another exception occurred: + +ProxyError Traceback (most recent call last) +Cell In[3], line 6 + 3 from zipfile import ZipFile + 5 # Download the dataset +----> 6 ck = requests.get('http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip').content + 7 zipped = ZipFile(BytesIO(ck)) + 8 df = pd.read_table( + 9 BytesIO(zipped.read('OxCodeIntroStateSpaceBook/Chapter_2/NorwayFinland.txt')), + 10 skiprows=1, header=None, sep='\s+', engine='python', + 11 names=['date','nf', 'ff'] + 12 ) + +File /usr/lib/python3/dist-packages/requests/api.py:73, in get(url, params, **kwargs) + 62 def get(url, params=None, **kwargs): + 63 r"""Sends a GET request. + 64 + 65 :param url: URL for the new :class:`Request` object. + (...) + 70 :rtype: requests.Response + 71 """ +---> 73 return request("get", url, params=params, **kwargs) + +File /usr/lib/python3/dist-packages/requests/api.py:59, in request(method, url, **kwargs) + 55 # By using the 'with' statement we are sure the session is closed, thus we + 56 # avoid leaving sockets open which can trigger a ResourceWarning in some + 57 # cases, and look like a memory leak in others. + 58 with sessions.Session() as session: +---> 59 return session.request(method=method, url=url, **kwargs) + +File /usr/lib/python3/dist-packages/requests/sessions.py:589, in Session.request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json) + 584 send_kwargs = { + 585 "timeout": timeout, + 586 "allow_redirects": allow_redirects, + 587 } + 588 send_kwargs.update(settings) +--> 589 resp = self.send(prep, **send_kwargs) + 591 return resp + +File /usr/lib/python3/dist-packages/requests/sessions.py:703, in Session.send(self, request, **kwargs) + 700 start = preferred_clock() + 702 # Send the request +--> 703 r = adapter.send(request, **kwargs) + 705 # Total elapsed time of the request (approximately) + 706 elapsed = preferred_clock() - start + +File /usr/lib/python3/dist-packages/requests/adapters.py:694, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) + 691 raise RetryError(e, request=request) + 693 if isinstance(e.reason, _ProxyError): +--> 694 raise ProxyError(e, request=request) + 696 if isinstance(e.reason, _SSLError): + 697 # This branch is for urllib3 v1.22 and later. + 698 raise SSLError(e, request=request) + +ProxyError: HTTPConnectionPool(host='127.0.0.1', port=9): Max retries exceeded with url: http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) + +An error occurred while executing the following cell: +------------------ +import requests +from io import BytesIO +from zipfile import ZipFile + +# Download the dataset +ck = requests.get('http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip').content +zipped = ZipFile(BytesIO(ck)) +df = pd.read_table( + BytesIO(zipped.read('OxCodeIntroStateSpaceBook/Chapter_2/NorwayFinland.txt')), + skiprows=1, header=None, sep='\s+', engine='python', + names=['date','nf', 'ff'] +) +------------------ + +----- stderr ----- +<>:10: SyntaxWarning: invalid escape sequence '\s' +<>:10: SyntaxWarning: invalid escape sequence '\s' +/tmp/ipykernel_3504114/2613092943.py:10: SyntaxWarning: invalid escape sequence '\s' + skiprows=1, header=None, sep='\s+', engine='python', +----- stderr ----- +/tmp/ipykernel_3504114/2613092943.py:10: SyntaxWarning: invalid escape sequence '\s' + skiprows=1, header=None, sep='\s+', engine='python', +------------------ + +--------------------------------------------------------------------------- +ConnectionRefusedError Traceback (most recent call last) +File /usr/lib/python3/dist-packages/urllib3/connection.py:198, in HTTPConnection._new_conn(self) + 197 try: +--> 198 sock = connection.create_connection( + 199 (self._dns_host, self.port), + 200 self.timeout, + 201 source_address=self.source_address, + 202 socket_options=self.socket_options, + 203 ) + 204 except socket.gaierror as e: + +File /usr/lib/python3/dist-packages/urllib3/util/connection.py:85, in create_connection(address, timeout, source_address, socket_options) + 84 try: +---> 85 raise err + 86 finally: + 87 # Break explicitly a reference cycle + +File /usr/lib/python3/dist-packages/urllib3/util/connection.py:73, in create_connection(address, timeout, source_address, socket_options) + 72 sock.bind(source_address) +---> 73 sock.connect(sa) + 74 # Break explicitly a reference cycle + +ConnectionRefusedError: [Errno 111] Connection refused + +The above exception was the direct cause of the following exception: + +NewConnectionError Traceback (most recent call last) +File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:787, in HTTPConnectionPool.urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw) + 786 # Make the request on the HTTPConnection object +--> 787 response = self._make_request( + 788 conn, + 789 method, + 790 url, + 791 timeout=timeout_obj, + 792 body=body, + 793 headers=headers, + 794 chunked=chunked, + 795 retries=retries, + 796 response_conn=response_conn, + 797 preload_content=preload_content, + 798 decode_content=decode_content, + 799 **response_kw, + 800 ) + 802 # Everything went great! + +File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:493, in HTTPConnectionPool._make_request(self, conn, method, url, body, headers, retries, timeout, chunked, response_conn, preload_content, decode_content, enforce_content_length) + 492 try: ****************************************************************************** ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_sarimax_internet.ipynb @@ -7492,10 +7783,10 @@ ----- stderr ----- <>:8: SyntaxWarning: invalid escape sequence '\s' <>:8: SyntaxWarning: invalid escape sequence '\s' -/tmp/ipykernel_1660927/1758367982.py:8: SyntaxWarning: invalid escape sequence '\s' +/tmp/ipykernel_3502547/1758367982.py:8: SyntaxWarning: invalid escape sequence '\s' skiprows=1, header=None, sep='\s+', engine='python', ----- stderr ----- -/tmp/ipykernel_1660927/1758367982.py:8: SyntaxWarning: invalid escape sequence '\s' +/tmp/ipykernel_3502547/1758367982.py:8: SyntaxWarning: invalid escape sequence '\s' skiprows=1, header=None, sep='\s+', engine='python', ------------------ @@ -7539,27 +7830,7 @@ File /usr/lib/python3.13/http/client.py:1001, in HTTPConnection.connect(self) 1000 sys.audit("http.client.connect", self, self.host, self.port) -> 1001 self.sock = self._create_connection( - 1002 (self. -At iterate 5 f= 2.23109D+00 |proj g|= 3.93605D-05 - - * * * - -Tit = total number of iterations -Tnf = total number of function evaluations -Tnint = total number of segments explored during Cauchy searches -Skip = number of BFGS updates skipped -Nact = number of active bounds at final generalized Cauchy point -Projg = norm of the final projected gradient -F = final function value - - * * * - - N Tit Tnf Tnint Skip Nact Projg F - 3 6 8 1 0 0 7.065D-07 2.231D+00 - F = 2.2310884444664767 - -CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL -host,self.port), self.timeout, self.source_address) + 1002 (self.host,self.port), self.timeout, self.source_address) 1003 # Might fail in OSs that don't implement TCP_NODELAY File /usr/lib/python3.13/socket.py:864, in create_connection(address, timeout, source_address, all_errors) @@ -7688,7 +7959,16 @@ File /usr/lib/python3.13/urllib/request.py:1322, in AbstractHTTPHandler.do_open(self, http_class, req, **http_conn_args) 1319 h.request(req.get_method(), req.selector, req.data, headers, 1320 encode_chunked=req.has_header('Transfer-encoding')) - 1321 except OSError as err: # timeout error +--> 493 conn.request( + 494 method, + 495 url, + 496 body=body, + 497 headers=headers, + 498 chunked=chunked, + 499 preload_content=preload_content, + 500 decode_content=decode_content 1321, + 501 enforce_content_length except =OSError as err: enforce_content_length# timeout error +, -> 1322 raise URLError(err) 1323 r = h.getresponse() 1324 except: @@ -7712,10 +7992,10 @@ ----- stderr ----- <>:8: SyntaxWarning: invalid escape sequence '\s' <>:8: SyntaxWarning: invalid escape sequence '\s' -/tmp/ipykernel_1660927/1758367982.py:8: SyntaxWarning: invalid escape sequence '\s' +/tmp/ipykernel_3502547/1758367982.py:8: SyntaxWarning: invalid escape sequence '\s' skiprows=1, header=None, sep='\s+', engine='python', ----- stderr ----- -/tmp/ipykernel_1660927/1758367982.py:8: SyntaxWarning: invalid escape sequence '\s' +/tmp/ipykernel_3502547/1758367982.py:8: SyntaxWarning: invalid escape sequence '\s' skiprows=1, header=None, sep='\s+', engine='python', ------------------ @@ -7724,35 +8004,175 @@ File /usr/lib/python3.13/urllib/request.py:1319, in AbstractHTTPHandler.do_open(self, http_class, req, **http_conn_args) 1318 try: -> 1319 h.request(req.get_method(), req.selector, req.data, headers, - 1320 encode_chunked=req.has_header('Transfer-encoding')) - 1321 except OSError as err: # timeout error + 502 ) + 504 # We are swallowing BrokenPipeError (errno.EPIPE) since the server is 1320 +encode_chunked= 505 req.# legitimately able to close the connection after sending a valid response.has_header(' + 506Transfer-encoding # With this behaviour, the received response is still readable. + +File /usr/lib/python3/dist-packages/urllib3/connection.py:445, in ')HTTPConnection.request(self, method, url, body, headers, chunked, preload_content, decode_content, enforce_content_length) + 444 self) +.putheader(header, value) +--> 445 self. 1321endheaders() + 447 # If we're given a body we start sending that in chunks. + +File /usr/lib/python3.13/http/client.py:1331, in HTTPConnection.endheadersexcept (self, message_body, encode_chunked) + 1330 raiseOSError CannotSendHeader() +-> 1331 self. as err: # timeout error File /usr/lib/python3.13/http/client.py:1336, in HTTPConnection.request(self, method, url, body, headers, encode_chunked) 1335 """Send a complete request to the server.""" --> 1336 self._send_request(method, url, body, headers, encode_chunked) +-> 1336 self._send_output_send_request(method,( urlmessage_body, body,, headers, encode_chunkedencode_chunked) -File /usr/lib/python3.13/http/client.py:1382, in HTTPConnection._send_request(self, method, url, body, headers, encode_chunked) - 1381 body = _encode(body, 'body') --> 1382 self.endheaders(body, encode_chunked=encode_chunked) - -File /usr/lib/python3.13/http/client.py:1331, in HTTPConnection.endheaders(self, message_body, encode_chunked) - 1330 raise CannotSendHeader() --> 1331 self._send_output(message_body, encode_chunked=encode_chunked) +File =/usr/lib/python3.13/http/client.py:1382, in encode_chunkedHTTPConnection._send_request(self, method, url, body, headers, encode_chunked) + 1381) body = -File /usr/lib/python3.13/http/client.py:1091, in HTTPConnection._send_output(self, message_body, encode_chunked) - 1090 del self._buffer[:] --> 1091 self.send(msg) - 1093 if message_body is not None: +File _encode(body, '/usr/lib/python3.13/http/client.py:1091body', in ) +-> 1382HTTPConnection._send_output self(self, message_body, encode_chunked).endheaders + 1090(body , encode_chunked=delencode_chunked) + +File /usr/lib/python3.13/http/client.py:1331self, in HTTPConnection.endheaders.(self, message_body, encode_chunked) + 1330 _buffer[:] +-> 1091raise CannotSendHeader() + -> 1331 selfself._send_output(.message_body,send encode_chunked(msg=encode_chunked) + +File )/usr/lib/python3.13/http/client.py:1091 + 1093, in HTTPConnection._send_output (self, message_body, encode_chunked) + 1090 ifdel message_body self.is_buffer[:] +-> 1091 self .notsend(msg )None: 1094 1095 # create a consistent interface to message_body -File /usr/lib/python3.13/http/client.py:1035, in HTTPConnection.send(self, data) - 1034 if self.auto_open: --> 1035 self.connect() - 1036 else: -File /usr/lib/python3.13/http/client.py:1470, in HTTPSConnection.connect(self) - 1468 "Connect to a host on a given (SSL) port." +File /usr/lib/python3.13/http/client.py:1035, in HTTPConnection.send(self, data) 1093 + 1034 if ifself.auto_open: +-> 1035 message_body is self.connect() + 1036 not None: +else 1094 + 1095 : + +File # create a consistent interface to message_body + +File /usr/lib/python3.13/http/client.py:1035/usr/lib/python3/dist-packages/urllib3/connection.py:276, in HTTPConnection.send, in (self, data) +HTTPConnection.connect 1034 (self)if +self. 275auto_open: +-> 1035 def connect(self) ->self. None: +--> 276 self.sock = self.connect_new_conn() +( 277 if self.)_tunnel_host: + 278 # If we're tunneling it means we're connected to our proxy. + +File /usr/lib/python3/dist-packages/urllib3/connection.py:213, in + 1036HTTPConnection._new_conn(self) + 212 except OSError as e: +else: + +File --> 213 raise NewConnectionError( +/usr/lib/python3.13/http/client.py:1470 214 self, f", in Failed to establish a new connection: HTTPSConnection.connect{e}" + 215(self) ) from e + 217 + sys.audit("http.client.connect", self, self. 1468host, self.port) + + "NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused + +The above exception was the direct cause of the following exception: + +ProxyError Traceback (most recent call last) +ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) + +The above exception was the direct cause of the following exception: + +MaxRetryError Traceback (most recent call last) +File /usr/lib/python3/dist-packages/requests/adapters.py:667, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) + 666 try: +--> 667 resp = conn.urlopen( + 668 method=request.method, + 669 url=url, + 670 body=request.body, + 671 headers=request.headers, + 672 redirect=False, + 673 assert_same_host=False, + 674 preload_content=False, + 675 decode_content=False, + 676 retries=self.max_retries, + 677 timeout=timeout, + 678 chunked=chunked, + 679 ) + 681 except (ProtocolError, OSError) as err: + +File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:841, in HTTPConnectionPool.urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw) + 839 new_e = ProtocolError("Connection aborted.", new_e) +--> 841 retries = retries.increment( + 842 method, url, error=new_e, _pool=self, _stacktrace=sys.exc_info()[2] + 843 ) + 844 retries.sleep() + +File /usr/lib/python3/dist-packages/urllib3/util/retry.py:519, in Retry.increment(self, method, url, response, error, _pool, _stacktrace) + 518 reason = error or ResponseError(cause) +--> 519 raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] + 521 log.debug("Incremented Retry for (url='%s'): %r", url, new_retry) + +MaxRetryError: HTTPConnectionPool(host='127.0.0.1', port=9): Max retries exceeded with url: http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) + +During handling of the above exception, another exception occurred: + +ProxyError Traceback (most recent call last) +Cell In[3], line 6 + 3 from zipfile import ZipFile + 5 # Download the dataset +----> 6 ck = requests.get('http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip').content + 7 zipped = ZipFile(BytesIO(ck)) + 8 df = pd.read_table( + 9 BytesIO(zipped.read('OxCodeIntroStateSpaceBook/Chapter_2/NorwayFinland.txt')), + 10 skiprows=1, header=None, sep='\s+', engine='python', + 11 names=['date','nf', 'ff'] + 12 ) + +File /usr/lib/python3/dist-packages/requests/api.py:73, in get(url, params, **kwargs) + 62 def get(url, params=None, **kwargs): + 63 r"""Sends a GET request. + 64 + 65 :param url: URL for the new :class:`Request` object. + (...) + 70 :rtype: requests.Response + 71 """ +---> 73 return request("get", url, params=params, **kwargs) + +File /usr/lib/python3/dist-packages/requests/api.py:59, in request(method, url, **kwargs) + 55 # By using the 'with' statement we are sure the session is closed, thus we + 56 # avoid leaving sockets open which can trigger a ResourceWarning in some + 57 # cases, and look like a memory leak in others. + 58 with sessions.Session() as session: +---> 59 return session.request(method=method, url=url, **kwargs) + +File /usr/lib/python3/dist-packages/requests/sessions.py:589, in Session.request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json) + 584 send_kwargs = { + 585 "timeout": timeout, + 586 "allow_redirects": allow_redirects, + 587 } + 588 send_kwargs.update(settings) +--> 589 resp = self.send(prep, **send_kwargs) + 591 return resp + +File /usr/lib/python3/dist-packages/requests/sessions.py:703, in Session.send(self, request, **kwargs) + 700 start = preferred_clock() + 702 # Send the request +--> 703 r = adapter.send(request, **kwargs) + 705 # Total elapsed time of the request (approximately) + 706 elapsed = preferred_clock() - start + +File /usr/lib/python3/dist-packages/requests/adapters.py:694, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) + 691 raise RetryError(e, request=request) + 693 if isinstance(e.reason, _ProxyError): +--> 694 raise ProxyError(e, request=request) + 696 if isinstance(e.reason, _SSLError): + 697 # This branch is for urllib3 v1.22 and later. + 698 raise SSLError(e, request=request) + +ProxyError: HTTPConnectionPool(host='127.0.0.1', port=9): Max retries exceeded with url: http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) + +****************************************************************************** + + +Connect to a host on a given (SSL) port." -> 1470 super().connect() 1472 if self._tunnel_host: @@ -7796,7 +8216,8 @@ 1403 kwds.update(kwds_defaults) -> 1405 return _read(filepath_or_buffer, kwds) -File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:620, in _read(filepath_or_buffer, kwds) +File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:620Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_dfm_coincident.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_dfm_coincident.ipynb +, in _read(filepath_or_buffer, kwds) 617 _validate_names(kwds.get("names", None)) 619 # Create the parser. --> 620 parser = TextFileReader(filepath_or_buffer, **kwds) @@ -7876,7 +8297,9 @@ File /usr/lib/python3.13/urllib/request.py:466, in OpenerDirector._call_chain(self, chain, kind, meth_name, *args) 464 for handler in handlers: 465 func = getattr(handler, meth_name) ---> 466 result = func(*args) +--> 466 result = func +At iterate 10 f= 1.41329D+00 |proj g|= 3.11658D-05 +(*args) 467 if result is not None: 468 return result @@ -7898,71 +8321,7 @@ ****************************************************************************** -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_dfm_coincident.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_dfm_coincident.ipynb -RUNNING THE L-BFGS-B CODE - - * * * - -Machine precision = 2.220D-16 - N = 3 M = 10 - This problem is unconstrained. - -At X0 0 variables are exactly at the bounds - -At iterate 0 f= 2.22839D+00 |proj g|= 2.38554D-03 - -At iterate 5 f= 2.22838D+00 |proj g|= 9.78773D-08 - - * * * - -Tit = total number of iterations -Tnf = total number of function evaluations -Tnint = total number of segments explored during Cauchy searches -Skip = number of BFGS updates skipped -Nact = number of active bounds at final generalized Cauchy point -Projg = norm of the final projected gradient -F = final function value - - * * * - - N Tit Tnf Tnint Skip Nact Projg F - 3 5 8 1 0 0 9.788D-08 2.228D+00 - F = 2.2283821699856383 - -CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL -RUNNING THE L-BFGS-B CODE - - * * * - -Machine precision = 2.220D-16 - N = 3 M = 10 - This problem is unconstrained. - -At X0 0 variables are exactly at the bounds - -At iterate 0 f= 2.23132D+00 |proj g|= 1.09171D-02 - -At iterate 5 f= 2.23109D+00 |proj g|= 3.93605D-05 - - * * * - -Tit = total number of iterations -Tnf = total number of function evaluations -Tnint = total number of segments explored during Cauchy searches -Skip = number of BFGS updates skipped -Nact = number of active bounds at final generalized Cauchy point -Projg = norm of the final projected gradient -F = final function value - - * * * - - N Tit Tnf Tnint Skip Nact Projg F - 3 6 8 1 0 0 7.065D-07 2.231D+00 - F = 2.2310884444664767 - -CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL - -At iterate 10 f= 1.41329D+00 |proj g|= 3.11658D-05 +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_cycles.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_cycles.ipynb * * * @@ -7991,9 +8350,9 @@ At X0 0 variables are exactly at the bounds -At iterate 0 f= 2.23132D+00 |proj g|= 1.09171D-02 +At iterate 0 f= 2.32873D+00 |proj g|= 8.23649D-03 -At iterate 5 f= 2.23109D+00 |proj g|= 3.93605D-05 +At iterate 5 f= 2.32864D+00 |proj g|= 1.41994D-03 * * * @@ -8008,21 +8367,10 @@ * * * N Tit Tnf Tnint Skip Nact Projg F - 3 6 8 1 0 0 7.065D-07 2.231D+00 - F = 2.2310884444664767 + 3 8 10 1 0 0 5.820D-06 2.329D+00 + F = 2.3286389358138648 CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL -RUNNING THE L-BFGS-B CODE - - * * * - -Machine precision = 2.220D-16 - N = 3 M = 10 - This problem is unconstrained. - -At X0 0 variables are exactly at the bounds - -At iterate 0 f= 2.58518D+00 |proj g|= 5.99456D-05 ****************************************************************************** ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_news.ipynb @@ -8071,7 +8419,20 @@ ****************************************************************************** -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_cycles.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_cycles.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_concentrated_scale.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_concentrated_scale.ipynb +RUNNING THE L-BFGS-B CODE + + * * * + +Machine precision = 2.220D-16 + N = 3 M = 10 + This problem is unconstrained. + +At X0 0 variables are exactly at the bounds + +At iterate 0 f= 2.23132D+00 |proj g|= 1.09171D-02 + +At iterate 5 f= 2.23109D+00 |proj g|= 3.93605D-05 * * * @@ -8086,470 +8447,65 @@ * * * N Tit Tnf Tnint Skip Nact Projg F - 3 3 5 1 0 0 3.347D-05 2.585D+00 - F = 2.5851830060985592 - -CONVERGENCE: REL_REDUCTION_OF_F_<=_FACTR*EPSMCH - -****************************************************************************** -ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_local_linear_trend.ipynb -An error occurred while executing the following cell: ------------------- -import requests -from io import BytesIO -from zipfile import ZipFile - -# Download the dataset -ck = requests.get('http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip').content -zipped = ZipFile(BytesIO(ck)) -df = pd.read_table( - BytesIO(zipped.read('OxCodeIntroStateSpaceBook/Chapter_2/NorwayFinland.txt')), - skiprows=1, header=None, sep='\s+', engine='python', - names=['date','nf', 'ff'] -) ------------------- - ------ stderr ----- -<>:10: SyntaxWarning: invalid escape sequence '\s' -<>:10: SyntaxWarning: invalid escape sequence '\s' -/tmp/ipykernel_1661302/2613092943.py:10: SyntaxWarning: invalid escape sequence '\s' - skiprows=1, header=None, sep='\s+', engine='python', ------ stderr ----- -/tmp/ipykernel_1661302/2613092943.py:10: SyntaxWarning: invalid escape sequence '\s' - skiprows=1, header=None, sep='\s+', engine='python', ------------------- - ---------------------------------------------------------------------------- -ConnectionRefusedError Traceback (most recent call last) -File /usr/lib/python3/dist-packages/urllib3/connection.py:198, in HTTPConnection._new_conn(self) - 197 try: ---> 198 sock = connection.create_connection( - 199 (self._dns_host, self.port), - 200 self.timeout, - 201 source_address=self.source_address, - 202 socket_options=self.socket_options, - 203 ) - 204 except socket.gaierror as e: - -File /usr/lib/python3/dist-packages/urllib3/util/connection.py:85, in create_connection(address, timeout, source_address, socket_options) - 84 try: ----> 85 raise err - 86 finally: - 87 # Break explicitly a reference cycle - -File /usr/lib/python3/dist-packages/urllib3/util/connection.py:73, in create_connection(address, timeout, source_address, socket_options) - 72 sock.bind(source_address) ----> 73 sock.connect(sa) - 74 # Break explicitly a reference cycle - -ConnectionRefusedError: [Errno 111] Connection refused - -The above exception was the direct cause of the following exception: - -NewConnectionError Traceback (most recent call last) -File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:787, in HTTPConnectionPool.urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw) - 786 # Make the request on the HTTPConnection object ---> 787 response = self._make_request( - 788 conn, - 789 method, - 790 url, - 791 timeout=timeout_obj, - 792 body=body, - 793 headers=headers, - 794 chunked=chunked, - 795 retries=retries, - 796 response_conn=response_conn, - 797 preload_content=preload_content, - 798 decode_content=decode_content, - 799 **response_kw, - 800 ) - 802 # Everything went great! - -File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:493, in HTTPConnectionPool._make_request(self, conn, method, url, body, headers, retries, timeout, chunked, response_conn, preload_content, decode_content, enforce_content_length) - 492 try: ---> 493 conn.request( - 494 method, - 495 url, - 496 body=body, - 497 headers=headers, - 498 chunked=chunked, - 499 preload_content=preload_content, - 500 decode_content=decode_content, - 501 enforce_content_length=enforce_content_length, - 502 ) - 504 # We are swallowing BrokenPipeError (errno.EPIPE) since the server is - 505 # legitimately able to close the connection after sending a valid response. - 506 # With this behaviour, the received response is still readable. - -File /usr/lib/python3/dist-packages/urllib3/connection.py:445, in HTTPConnection.request(self, method, url, body, headers, chunked, preload_content, decode_content, enforce_content_length) - 444 self.putheader(header, value) ---> 445 self.endheaders() - 447 # If we're given a body we start sending that in chunks. - -File /usr/lib/python3.13/http/client.py:1331, in HTTPConnection.endheaders(self, message_body, encode_chunked) - 1330 raise CannotSendHeader() --> 1331 self._send_output(message_body, encode_chunked=encode_chunked) - -File /usr/lib/python3.13/http/client.py:1091, in HTTPConnection._send_output(self, message_body, encode_chunked) - 1090 del self._buffer[:] --> 1091 self.send(msg) - 1093 if message_body is not None: - 1094 - 1095 # create a consistent interface to message_body - -File /usr/lib/python3.13/http/client.py:1035, in HTTPConnection.send(self, data) - 1034 if self.auto_open: --> 1035 self.connect() - 1036 else: - -File /usr/lib/python3/dist-packages/urllib3/connection.py:276, in HTTPConnection.connect(self) - 275 def connect(self) -> None: ---> 276 self.sock = self._new_conn() - 277 if self._tunnel_host: - 278 # If we're tunneling it means we're connected to our proxy. - -File /usr/lib/python3/dist-packages/urllib3/connection.py:213, in HTTPConnection._new_conn(self) - 212 except OSError as e: ---> 213 raise NewConnectionError( - 214 self, f"Failed to establish a new connection: {e}" - 215 ) from e - 217 sys.audit("http.client.connect", self, self.host, self.port) - -NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused - -The above exception was the direct cause of the following exception: - -ProxyError Traceback (most recent call last) -ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) - -The above exception was the direct cause of the following exception: - -MaxRetryError Traceback (most recent call last) -File /usr/lib/python3/dist-packages/requests/adapters.py:667, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) - 666 try: ---> 667 resp = conn.urlopen( - 668 method=request.method, - 669 url=url, - 670 body=request.body, - 671 headers=request.headers, - 672 redirect=False, - 673 assert_same_host=False, - 674 preload_content=False, - 675 decode_content=False, - 676 retries=self.max_retries, - 677 timeout=timeout, - 678 chunked=chunked, - 679 ) - 681 except (ProtocolError, OSError) as err: - -File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:841, in HTTPConnectionPool.urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw) - 839 new_e = ProtocolError("Connection aborted.", new_e) ---> 841 retries = retries.increment( - 842 method, url, error=new_e, _pool=self, _stacktrace=sys.exc_info()[2] - 843 ) - 844 retries.sleep() - -File /usr/lib/python3/dist-packages/urllib3/util/retry.py:519, in Retry.increment(self, method, url, response, error, _pool, _stacktrace) - 518 reason = error or ResponseError(cause) ---> 519 raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] - 521 log.debug("Incremented Retry for (url='%s'): %r", url, new_retry) - -MaxRetryError: HTTPConnectionPool(host='127.0.0.1', port=9): Max retries exceeded with url: http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) - -During handling of the above exception, another exception occurred: - -ProxyError Traceback (most recent call last) -Cell In[3], line 6 - 3 from zipfile import ZipFile - 5 # Download the dataset -----> 6 ck = requests.get('http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip').content - 7 zipped = ZipFile(BytesIO(ck)) - 8 df = pd.read_table( - 9 BytesIO(zipped.read('OxCodeIntroStateSpaceBook/Chapter_2/NorwayFinland.txt')), - 10 skiprows=1, header=None, sep='\s+', engine='python', - 11 names=['date','nf', 'ff'] - 12 ) - -File /usr/lib/python3/dist-packages/requests/api.py:73, in get(url, params, **kwargs) - 62 def get(url, params=None, **kwargs): - 63 r"""Sends a GET request. - 64 - 65 :param url: URL for the new :class:`Request` object. - (...) - 70 :rtype: requests.Response - 71 """ ----> 73 return request("get", url, params=params, **kwargs) - -File /usr/lib/python3/dist-packages/requests/api.py:59, in request(method, url, **kwargs) - 55 # By using the 'with' statement we are sure the session is closed, thus we - 56 # avoid leaving sockets open which can trigger a ResourceWarning in some - 57 # cases, and look like a memory leak in others. - 58 with sessions.Session() as session: ----> 59 return session.request(method=method, url=url, **kwargs) - -File /usr/lib/python3/dist-packages/requests/sessions.py:589, in Session.request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json) - 584 send_kwargs = { - 585 "timeout": timeout, - 586 "allow_redirects": allow_redirects, - 587 } - 588 send_kwargs.update(settings) ---> 589 resp = self.send(prep, **send_kwargs) - 591 return resp - -File /usr/lib/python3/dist-packages/requests/sessions.py:703, in Session.send(self, request, **kwargs) - 700 start = preferred_clock() - 702 # Send the request ---> 703 r = adapter.send(request, **kwargs) - 705 # Total elapsed time of the request (approximately) - 706 elapsed = preferred_clock() - start - -File /usr/lib/python3/dist-packages/requests/adapters.py:694, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) - 691 raise RetryError(e, request=request) - 693 if isinstance(e.reason, _ProxyError): ---> 694 raise ProxyError(e, request=request) - 696 if isinstance(e.reason, _SSLError): - 697 # This branch is for urllib3 v1.22 and later. - 698 raise SSLError(e, request=request) - -ProxyError: HTTPConnectionPool(host='127.0.0.1', port=9): Max retries exceeded with url: http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) - -An error occurred while executing the following cell: ------------------- -import requests -from io import BytesIO -from zipfile import ZipFile - -# Download the dataset -ck = requests.get('http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip').content -zipped = ZipFile(BytesIO(ck)) -df = pd.read_table( - BytesIO(zipped.read('OxCodeIntroStateSpaceBook/Chapter_2/NorwayFinland.txt')), - skiprows=1, header=None, sep='\s+', engine='python', - names=['date','nf', 'ff'] -) ------------------- - ------ stderr ----- -<>:10: SyntaxWarning: invalid escape sequence '\s' -<>:10: SyntaxWarning: invalid escape sequence '\s' -/tmp/ipykernel_1661302/2613092943.py:10: SyntaxWarning: invalid escape sequence '\s' - skiprows=1, header=None, sep='\s+', engine='python', ------ stderr ----- -/tmp/ipykernel_1661302/2613092943.py:10: SyntaxWarning: invalid escape sequence '\s' - skiprows=1, header=None, sep='\s+', engine='python', ------------------- - ---------------------------------------------------------------------------- -ConnectionRefusedError Traceback (most recent call last) -File /usr/lib/python3/dist-packages/urllib3/connection.py:198, in HTTPConnection._new_conn(self) - 197 try: ---> 198 sock = connection.create_connection( - 199 (self._dns_host, self.port), - 200 self.timeout, - 201 source_address=self.source_address, - 202 socket_options=self.socket_options, - 203 ) - 204 except socket.gaierror as e: - -File /usr/lib/python3/dist-packages/urllib3/util/connection.py:85, in create_connection(address, timeout, source_address, socket_options) - 84 try: ----> 85 raise err - 86 finally: - 87 # Break explicitly a reference cycle - -File /usr/lib/python3/dist-packages/urllib3/util/connection.py:73, in create_connection(address, timeout, source_address, socket_options) - 72 sock.bind(source_address) ----> 73 sock.connect(sa) - 74 # Break explicitly a reference cycle - -ConnectionRefusedError: [Errno 111] Connection refused - -The above exception was the direct cause of the following exception: - -NewConnectionError Traceback (most recent call last) -File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:787, in HTTPConnectionPool.urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw) - 786 # Make the request on the HTTPConnection object ---> 787 response = self._make_request( - 788 conn, - 789 method, - 790 url, - 791 timeout=timeout_obj, - 792 body=body, - 793 headers=headers, - 794 chunked=chunked, - 795 retries=retries, - 796 response_conn=response_conn, - 797 preload_content=preload_content, - 798 decode_content=decode_content, - 799 **response_kw, - 800 ) - 802 # Everything went great! - -File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:493, in HTTPConnectionPool._make_request(self, conn, method, url, body, headers, retries, timeout, chunked, response_conn, preload_content, decode_content, enforce_content_length) - 492 try: ---> 493 conn.request( - 494 method, - 495 url, - 496 body=body, - 497 headers=headers, - 498 chunked=chunked, - 499 preload_content=preload_content, - 500 decode_content=decode_content, - 501 enforce_content_length=enforce_content_length, - 502 ) - 504 # We are swallowing BrokenPipeError (errno.EPIPE) since the server is - 505 # legitimately able to close the connection after sending a valid response. - 506 # With this behaviour, the received response is still readable. - -File /usr/lib/python3/dist-packages/urllib3/connection.py:445, in HTTPConnection.request(self, method, url, body, headers, chunked, preload_content, decode_content, enforce_content_length) - 444 self.putheader(header, value) ---> 445 self.endheaders() - 447 # If we're given a body we start sending that in chunks. - -File /usr/lib/python3.13/http/client.py:1331, in HTTPConnection.endheaders(self, message_body, encode_chunked) - 1330 raise CannotSendHeader() --> 1331 self._send_output(message_body, encode_chunked=encode_chunked) - -File /usr/lib/python3.13/http/client.py:1091, in HTTPConnection._send_output(self, message_body, encode_chunked) - 1090 del self._buffer[:] --> 1091 self.send(msg) - 1093 if message_body is not None: - 1094 - 1095 # create a consistent interface to message_body - -File /usr/lib/python3.13/http/client.py:1035, in HTTPConnection.send(self, data) - 1034 if self.auto_open: --> 1035 self.connect() - 1036 else: - -File /usr/lib/python3/dist-packages/urllib3/connection.py:276, in HTTPConnection.connect(self) - 275 def connect(self) -> None: ---> 276 self.sock = self._new_conn() - 277 if self._tunnel_host: - 278 # If we're tunneling it means we're connected to our proxy. - -File /usr/lib/python3/dist-packages/urllib3/connection.py:213, in HTTPConnection._new_conn(self) - 212 except OSError as e: ---> 213 raise NewConnectionError( - 214 self, f"Failed to establish a new connection: {e}" - 215 ) from e - 217 sys.audit("http.client.connect", self, self.host, self.port) - -NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused - -The above exception was the direct cause of the following exception: - -ProxyError Traceback (most recent call last) -ProxyError: ('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) - -The above exception was the direct cause of the following exception: + 3 6 8 1 0 0 7.065D-07 2.231D+00 + F = 2.2310884444664767 -MaxRetryError Traceback (most recent call last) -File /usr/lib/python3/dist-packages/requests/adapters.py:667, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) - 666 try: ---> 667 resp = conn.urlopen( - 668 method=request.method, - 669 url=url, - 670 body=request.body, - 671 headers=request.headers, - 672 redirect=False, - 673 assert_same_host=False, - 674 preload_content=False, - 675 decode_content=False, - 676 retries=self.max_retries, - 677 timeout=timeout, - 678 chunked=chunked, - 679 ) - 681 except (ProtocolError, OSError) as err: +CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL +RUNNING THE L-BFGS-B CODE -File /usr/lib/python3/dist-packages/urllib3/connectionpool.py:841, in HTTPConnectionPool.urlopen(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw) - 839 new_e = ProtocolError("Connection aborted.", new_e) ---> 841 retries = retries.increment( - 842 method, url, error=new_e, _pool=self, _stacktrace=sys.exc_info()[2] - 843 ) - 844 retries.sleep() + * * * -File /usr/lib/python3/dist-packages/urllib3/util/retry.py:519, in Retry.increment(self, method, url, response, error, _pool, _stacktrace) - 518 reason = error or ResponseError(cause) ---> 519 raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] - 521 log.debug("Incremented Retry for (url='%s'): %r", url, new_retry) +Machine precision = 2.220D-16 + N = 3 M = 10 + This problem is unconstrained. -MaxRetryError: HTTPConnectionPool(host='127.0.0.1', port=9): Max retries exceeded with url: http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) +At X0 0 variables are exactly at the bounds -During handling of the above exception, another exception occurred: +At iterate 0 f= 2.22839D+00 |proj g|= 2.38554D-03 +RUNNING THE L-BFGS-B CODE -ProxyError Traceback (most recent call last) -Cell In[3], line 6 - 3 from zipfile import ZipFile - 5 # Download the dataset -----> 6 ck = requests.get('http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip').content - 7 zipped = ZipFile(BytesIO(ck)) - 8 df = pd.read_table( - 9 BytesIO(zipped.read('OxCodeIntroStateSpaceBook/Chapter_2/NorwayFinland.txt')), - 10 skiprows=1, header=None, sep='\s+', engine='python', - 11 names=['date','nf', 'ff'] - 12 ) + * * * -File /usr/lib/python3/dist-packages/requests/api.py:73, in get(url, params, **kwargs) - 62 def get(url, params=None, **kwargs): - 63 r"""Sends a GET request. - 64 - 65 :param url: URL for the new :class:`Request` object. - (...) - 70 :rtype: requests.Response - 71 """ ----> 73 return request("get", url, params=params, **kwargs) +Machine precision = 2.220D-16 + N = 3 M = 10 + This problem is unconstrained. -File /usr/lib/python3/dist-packages/requests/api.py:59, in request(method, url, **kwargs) - 55 # By using the 'with' statement we are sure the session is closed, thus we - 56 # avoid leaving sockets open which can trigger a ResourceWarning in some - 57 # cases, and look like a memory leak in others. - 58 with sessions.Session() as session: ----> 59 return session.request(method=method, url=url, **kwargs) +At X0 0 variables are exactly at the bounds -File /usr/lib/python3/dist-packages/requests/sessions.py:589, in Session.request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json) - 584 send_kwargs = { - 585 "timeout": timeout, - 586 "allow_redirects": allow_redirects, - 587 } - 588 send_kwargs.update(settings) ---> 589 resp = self.send(prep, **send_kwargs) - 591 return resp +At iterate 0 f= 2.58518D+00 |proj g|= 5.99456D-05 -File /usr/lib/python3/dist-packages/requests/sessions.py:703, in Session.send(self, request, **kwargs) - 700 start = preferred_clock() - 702 # Send the request ---> 703 r = adapter.send(request, **kwargs) - 705 # Total elapsed time of the request (approximately) - 706 elapsed = preferred_clock() - start +At iterate 5 f= 2.22838D+00 |proj g|= 9.78773D-08 -File /usr/lib/python3/dist-packages/requests/adapters.py:694, in HTTPAdapter.send(self, request, stream, timeout, verify, cert, proxies) - 691 raise RetryError(e, request=request) - 693 if isinstance(e.reason, _ProxyError): ---> 694 raise ProxyError(e, request=request) - 696 if isinstance(e.reason, _SSLError): - 697 # This branch is for urllib3 v1.22 and later. - 698 raise SSLError(e, request=request) + * * * -ProxyError: HTTPConnectionPool(host='127.0.0.1', port=9): Max retries exceeded with url: http://staff.feweb.vu.nl/koopman/projects/ckbook/OxCodeAll.zip (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) +Tit = total number of iterations +Tnf = total number of function evaluations +Tnint = total number of segments explored during Cauchy searches +Skip = number of BFGS updates skipped +Nact = number of active bounds at final generalized Cauchy point +Projg = norm of the final projected gradient +F = final function value -****************************************************************************** + * * * + N Tit Tnf Tnint Skip Nact Projg F + 3 5 8 1 0 0 9.788D-08 2.228D+00 + F = 2.2283821699856383 -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_concentrated_scale.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_concentrated_scale.ipynb +CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL RUNNING THE L-BFGS-B CODE * * * Machine precision = 2.220D-16 - N = 2 M = 10 + N = 3 M = 10 This problem is unconstrained. At X0 0 variables are exactly at the bounds -At iterate 0 f= 1.37900D-01 |proj g|= 4.66940D-01 +At iterate 0 f= 2.23132D+00 |proj g|= 1.09171D-02 -At iterate 5 f= 1.32476D-01 |proj g|= 6.00136D-06 +At iterate 5 f= 2.23109D+00 |proj g|= 3.93605D-05 * * * @@ -8564,11 +8520,29 @@ * * * N Tit Tnf Tnint Skip Nact Projg F - 2 5 10 1 0 0 6.001D-06 1.325D-01 - F = 0.13247641992895676 + 3 6 8 1 0 0 7.065D-07 2.231D+00 + F = 2.2310884444664767 CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL + * * * + +Tit = total number of iterations +Tnf = total number of function evaluations +Tnint = total number of segments explored during Cauchy searches +Skip = number of BFGS updates skipped +Nact = number of active bounds at final generalized Cauchy point +Projg = norm of the final projected gradient +F = final function value + + * * * + + N Tit Tnf Tnint Skip Nact Projg F + 3 3 5 1 0 0 3.347D-05 2.585D+00 + F = 2.5851830060985592 + +CONVERGENCE: REL_REDUCTION_OF_F_<=_FACTR*EPSMCH + ****************************************************************************** ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_fixed_params.ipynb An error occurred while executing the following cell: @@ -8621,6 +8595,68 @@ Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_chandrasekhar.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_chandrasekhar.ipynb +RUNNING THE L-BFGS-B CODE + + * * * + +Machine precision = 2.220D-16 + N = 3 M = 10 + This problem is unconstrained. + +At X0 0 variables are exactly at the bounds + +At iterate 0 f= 2.23132D+00 |proj g|= 1.09171D-02 + +At iterate 5 f= 2.23109D+00 |proj g|= 3.93605D-05 + + * * * + +Tit = total number of iterations +Tnf = total number of function evaluations +Tnint = total number of segments explored during Cauchy searches +Skip = number of BFGS updates skipped +Nact = number of active bounds at final generalized Cauchy point +Projg = norm of the final projected gradient +F = final function value + + * * * + + N Tit Tnf Tnint Skip Nact Projg F + 3 6 8 1 0 0 7.065D-07 2.231D+00 + F = 2.2310884444664767 + +CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL +RUNNING THE L-BFGS-B CODE + + * * * + +Machine precision = 2.220D-16 + N = 2 M = 10 + This problem is unconstrained. + +At X0 0 variables are exactly at the bounds + +At iterate 0 f= 1.37900D-01 |proj g|= 4.66940D-01 + +At iterate 5 f= 1.32476D-01 |proj g|= 6.00136D-06 + + * * * + +Tit = total number of iterations +Tnf = total number of function evaluations +Tnint = total number of segments explored during Cauchy searches +Skip = number of BFGS updates skipped +Nact = number of active bounds at final generalized Cauchy point +Projg = norm of the final projected gradient +F = final function value + + * * * + + N Tit Tnf Tnint Skip Nact Projg F + 2 5 10 1 0 0 6.001D-06 1.325D-01 + F = 0.13247641992895676 + +CONVERGENCE: NORM_OF_PROJECTED_GRADIENT_<=_PGTOL Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_arma_0.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/statespace_arma_0.ipynb ****************************************************************************** @@ -8679,8 +8715,6 @@ Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/rolling_ls.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/rolling_ls.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/robust_models_1.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/robust_models_1.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/regression_plots.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/regression_plots.ipynb ****************************************************************************** ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/statespace_cycles.ipynb @@ -8721,6 +8755,8 @@ ****************************************************************************** +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/robust_models_1.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/robust_models_1.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/regression_plots.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/regression_plots.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/recursive_ls.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/recursive_ls.ipynb ****************************************************************************** @@ -8836,7 +8872,6 @@ Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/quantile_regression.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/quantile_regression.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/postestimation_poisson.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/postestimation_poisson.ipynb ****************************************************************************** ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/recursive_ls.ipynb @@ -8889,6 +8924,7 @@ ****************************************************************************** +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/postestimation_poisson.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/postestimation_poisson.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/pca_fertility_factors.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/pca_fertility_factors.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/ordinal_regression.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/ordinal_regression.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/ols.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/ols.ipynb @@ -9269,8 +9305,6 @@ Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/lowess.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/lowess.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/linear_regression_diagnostics_plots.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/linear_regression_diagnostics_plots.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/kernel_density.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/kernel_density.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/interactions_anova.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/interactions_anova.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/influence_glm_logit.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/influence_glm_logit.ipynb ****************************************************************************** ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/markov_autoregression.ipynb @@ -9345,6 +9379,8 @@ ****************************************************************************** +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/interactions_anova.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/interactions_anova.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/influence_glm_logit.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/influence_glm_logit.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/gls.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/gls.ipynb ****************************************************************************** @@ -9718,103 +9754,60 @@ Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/glm_weights.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/glm_weights.ipynb Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/glm_formula.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/glm_formula.ipynb -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/glm.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/glm.ipynb ****************************************************************************** -ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/mstl_decomposition.ipynb +ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/interactions_anova.ipynb An error occurred while executing the following cell: ------------------ -url = "https://raw.githubusercontent.com/tidyverts/tsibbledata/master/data-raw/vic_elec/VIC2015/demand.csv" -df = pd.read_csv(url) ------------------- - - ---------------------------------------------------------------------------- -ConnectionRefusedError Traceback (most recent call last) -File /usr/lib/python3.13/urllib/request.py:1319, in AbstractHTTPHandler.do_open(self, http_class, req, **http_conn_args) - 1318 try: --> 1319 h.request(req.get_method(), req.selector, req.data, headers, - 1320 encode_chunked=req.has_header('Transfer-encoding')) - 1321 except OSError as err: # timeout error - -File /usr/lib/python3.13/http/client.py:1336, in HTTPConnection.request(self, method, url, body, headers, encode_chunked) - 1335 """Send a complete request to the server.""" --> 1336 self._send_request(method, url, body, headers, encode_chunked) - -File /usr/lib/python3.13/http/client.py:1382, in HTTPConnection._send_request(self, method, url, body, headers, encode_chunked) - 1381 body = _encode(body, 'body') --> 1382 self.endheaders(body, encode_chunked=encode_chunked) - -File /usr/lib/python3.13/http/client.py:1331, in HTTPConnection.endheaders(self, message_body, encode_chunked) - 1330 raise CannotSendHeader() --> 1331 self._send_output(message_body, encode_chunked=encode_chunked) - -File /usr/lib/python3.13/http/client.py:1091, in HTTPConnection._send_output(self, message_body, encode_chunked) - 1090 del self._buffer[:] --> 1091 self.send(msg) - 1093 if message_body is not None: - 1094 - 1095 # create a consistent interface to message_body - -File /usr/lib/python3.13/http/client.py:1035, in HTTPConnection.send(self, data) - 1034 if self.auto_open: --> 1035 self.connect() - 1036 else: +from urllib.request import urlopen +import numpy as np -File /usr/lib/python3.13/http/client.py:1470, in HTTPSConnection.connect(self) - 1468 "Connect to a host on a given (SSL) port." --> 1470 super().connect() - 1472 if self._tunnel_host: +np.set_printoptions(precision=4, suppress=True) -File /usr/lib/python3.13/http/client.py:1001, in HTTPConnection.connect(self) - 1000 sys.audit("http.client.connect", self, self.host, self.port) --> 1001 self.sock = self._create_connection( - 1002 (self.host,self.port), self.timeout, self.source_address) - 1003 # Might fail in OSs that don't implement TCP_NODELAY +import pandas as pd -File /usr/lib/python3.13/socket.py:864, in create_connection(address, timeout, source_address, all_errors) - 863 if not all_errors: ---> 864 raise exceptions[0] - 865 raise ExceptionGroup("create_connection failed", exceptions) +pd.set_option("display.width", 100) +import matplotlib.pyplot as plt +from statsmodels.formula.api import ols +from statsmodels.graphics.api import interaction_plot, abline_plot +from statsmodels.stats.anova import anova_lm -File /usr/lib/python3.13/socket.py:849, in create_connection(address, timeout, source_address, all_errors) - 848 sock.bind(source_address) ---> 849 sock.connect(sa) - 850 # Break explicitly a reference cycle +try: + salary_table = pd.read_csv("salary.table") +except: # recent pandas can read URL without urlopen + url = "http://stats191.stanford.edu/data/salary.table" + fh = urlopen(url) + salary_table = pd.read_table(fh) + salary_table.to_csv("salary.table") -ConnectionRefusedError: [Errno 111] Connection refused +E = salary_table.E +M = salary_table.M +X = salary_table.X +S = salary_table.S +------------------ -During handling of the above exception, another exception occurred: -URLError Traceback (most recent call last) -Cell In[10], line 2 - 1 url = "https://raw.githubusercontent.com/tidyverts/tsibbledata/master/data-raw/vic_elec/VIC2015/demand.csv" -----> 2 df = pd.read_csv(url) +--------------------------------------------------------------------------- +FileNotFoundError Traceback (most recent call last) +Cell In[2], line 15 + 14 try: +---> 15 salary_table = pd.read_csv("salary.table") + 16 except: # recent pandas can read URL without urlopen File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1026, in read_csv(filepath_or_buffer, sep, delimiter, header, names, index_col, usecols, dtype, engine, converters, true_values, false_values, skipinitialspace, skiprows, skipfooter, nrows, na_values, keep_default_na, na_filter, verbose, skip_blank_lines, parse_dates, infer_datetime_format, keep_date_col, date_parser, date_format, dayfirst, cache_dates, iterator, chunksize, compression, thousands, decimal, lineterminator, quotechar, quoting, doublequote, escapechar, comment, encoding, encoding_errors, dialect, on_bad_lines, delim_whitespace, low_memory, memory_map, float_precision, storage_options, dtype_backend) - 1013 kwds_defaults = _refine_defaults_read( - 1014 dialect, - 1015 delimiter, - (...) - 1022 dtype_backend=dtype_backend, - 1023 ) 1024 kwds.update(kwds_defaults) -> 1026 return _read(filepath_or_buffer, kwds) File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:620, in _read(filepath_or_buffer, kwds) - 617 _validate_names(kwds.get("names", None)) 619 # Create the parser. --> 620 parser = TextFileReader(filepath_or_buffer, **kwds) 622 if chunksize or iterator: - 623 return parser File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1620, in TextFileReader.__init__(self, f, engine, **kwds) - 1617 self.options["has_index_names"] = kwds["has_index_names"] 1619 self.handles: IOHandles | None = None -> 1620 self._engine = self._make_engine(f, self.engine) File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1880, in TextFileReader._make_engine(self, f, engine) - 1878 if "b" not in mode: 1879 mode += "b" -> 1880 self.handles = get_handle( 1881 f, @@ -9827,87 +9820,24 @@ 1888 storage_options=self.options.get("storage_options", None), 1889 ) 1890 assert self.handles is not None - 1891 f = self.handles.handle - -File /usr/lib/python3/dist-packages/pandas/io/common.py:728, in get_handle(path_or_buf, mode, encoding, compression, memory_map, is_text, errors, storage_options) - 725 codecs.lookup_error(errors) - 727 # open URLs ---> 728 ioargs = _get_filepath_or_buffer( - 729 path_or_buf, - 730 encoding=encoding, - 731 compression=compression, - 732 mode=mode, - 733 storage_options=storage_options, - 734 ) - 736 handle = ioargs.filepath_or_buffer - 737 handles: list[BaseBuffer] -File /usr/lib/python3/dist-packages/pandas/io/common.py:384, in _get_filepath_or_buffer(filepath_or_buffer, encoding, compression, mode, storage_options) - 382 # assuming storage_options is to be interpreted as headers - 383 req_info = urllib.request.Request(filepath_or_buffer, headers=storage_options) ---> 384 with urlopen(req_info) as req: - 385 content_encoding = req.headers.get("Content-Encoding", None) - 386 if content_encoding == "gzip": - 387 # Override compression based on Content-Encoding header - -File /usr/lib/python3/dist-packages/pandas/io/common.py:289, in urlopen(*args, **kwargs) - 283 """ - 284 Lazy-import wrapper for stdlib urlopen, as that imports a big chunk of - 285 the stdlib. - 286 """ - 287 import urllib.request ---> 289 return urllib.request.urlopen(*args, **kwargs) - -File /usr/lib/python3.13/urllib/request.py:189, in urlopen(url, data, timeout, context) - 187 else: - 188 opener = _opener ---> 189 return opener.open(url, data, timeout) - -File /usr/lib/python3.13/urllib/request.py:489, in OpenerDirector.open(self, fullurl, data, timeout) - 486 req = meth(req) - 488 sys.audit('urllib.Request', req.full_url, req.data, req.headers, req.get_method()) ---> 489 response = self._open(req, data) - 491 # post-process response - 492 meth_name = protocol+"_response" - -File /usr/lib/python3.13/urllib/request.py:506, in OpenerDirector._open(self, req, data) - 503 return result - 505 protocol = req.type ---> 506 result = self._call_chain(self.handle_open, protocol, protocol + - 507 '_open', req) - 508 if result: - 509 return result - -File /usr/lib/python3.13/urllib/request.py:466, in OpenerDirector._call_chain(self, chain, kind, meth_name, *args) - 464 for handler in handlers: - 465 func = getattr(handler, meth_name) ---> 466 result = func(*args) - 467 if result is not None: - 468 return result - -File /usr/lib/python3.13/urllib/request.py:1367, in HTTPSHandler.https_open(self, req) - 1366 def https_open(self, req): --> 1367 return self.do_open(http.client.HTTPSConnection, req, - 1368 context=self._context) - -File /usr/lib/python3.13/urllib/request.py:1322, in AbstractHTTPHandler.do_open(self, http_class, req, **http_conn_args) - 1319 h.request(req.get_method(), req.selector, req.data, headers, - 1320 encode_chunked=req.has_header('Transfer-encoding')) - 1321 except OSError as err: # timeout error --> 1322 raise URLError(err) - 1323 r = h.getresponse() - 1324 except: - -URLError: +File /usr/lib/python3/dist-packages/pandas/io/common.py:873, in get_handle(path_or_buf, mode, encoding, compression, memory_map, is_text, errors, storage_options) + 871 if ioargs.encoding and "b" not in ioargs.mode: + 872 # Encoding +--> 873 handle = open( + 874 handle, + 875 ioargs.mode, + 876 encoding=ioargs.encoding, + 877 errors=errors, + 878 newline="", + 879 ) + 880 else: + 881 # Binary mode -An error occurred while executing the following cell: ------------------- -url = "https://raw.githubusercontent.com/tidyverts/tsibbledata/master/data-raw/vic_elec/VIC2015/demand.csv" -df = pd.read_csv(url) ------------------- +FileNotFoundError: [Errno 2] No such file or directory: 'salary.table' +During handling of the above exception, another exception occurred: ---------------------------------------------------------------------------- ConnectionRefusedError Traceback (most recent call last) File /usr/lib/python3.13/urllib/request.py:1319, in AbstractHTTPHandler.do_open(self, http_class, req, **http_conn_args) 1318 try: @@ -9939,11 +9869,6 @@ -> 1035 self.connect() 1036 else: -File /usr/lib/python3.13/http/client.py:1470, in HTTPSConnection.connect(self) - 1468 "Connect to a host on a given (SSL) port." --> 1470 super().connect() - 1472 if self._tunnel_host: - File /usr/lib/python3.13/http/client.py:1001, in HTTPConnection.connect(self) 1000 sys.audit("http.client.connect", self, self.host, self.port) -> 1001 self.sock = self._create_connection( @@ -9965,76 +9890,12 @@ During handling of the above exception, another exception occurred: URLError Traceback (most recent call last) -Cell In[10], line 2 - 1 url = "https://raw.githubusercontent.com/tidyverts/tsibbledata/master/data-raw/vic_elec/VIC2015/demand.csv" -----> 2 df = pd.read_csv(url) - -File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1026, in read_csv(filepath_or_buffer, sep, delimiter, header, names, index_col, usecols, dtype, engine, converters, true_values, false_values, skipinitialspace, skiprows, skipfooter, nrows, na_values, keep_default_na, na_filter, verbose, skip_blank_lines, parse_dates, infer_datetime_format, keep_date_col, date_parser, date_format, dayfirst, cache_dates, iterator, chunksize, compression, thousands, decimal, lineterminator, quotechar, quoting, doublequote, escapechar, comment, encoding, encoding_errors, dialect, on_bad_lines, delim_whitespace, low_memory, memory_map, float_precision, storage_options, dtype_backend) - 1013 kwds_defaults = _refine_defaults_read( - 1014 dialect, - 1015 delimiter, - (...) - 1022 dtype_backend=dtype_backend, - 1023 ) - 1024 kwds.update(kwds_defaults) --> 1026 return _read(filepath_or_buffer, kwds) - -File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:620, in _read(filepath_or_buffer, kwds) - 617 _validate_names(kwds.get("names", None)) - 619 # Create the parser. ---> 620 parser = TextFileReader(filepath_or_buffer, **kwds) - 622 if chunksize or iterator: - 623 return parser - -File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1620, in TextFileReader.__init__(self, f, engine, **kwds) - 1617 self.options["has_index_names"] = kwds["has_index_names"] - 1619 self.handles: IOHandles | None = None --> 1620 self._engine = self._make_engine(f, self.engine) - -File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1880, in TextFileReader._make_engine(self, f, engine) - 1878 if "b" not in mode: - 1879 mode += "b" --> 1880 self.handles = get_handle( - 1881 f, - 1882 mode, - 1883 encoding=self.options.get("encoding", None), - 1884 compression=self.options.get("compression", None), - 1885 memory_map=self.options.get("memory_map", False), - 1886 is_text=is_text, - 1887 errors=self.options.get("encoding_errors", "strict"), - 1888 storage_options=self.options.get("storage_options", None), - 1889 ) - 1890 assert self.handles is not None - 1891 f = self.handles.handle - -File /usr/lib/python3/dist-packages/pandas/io/common.py:728, in get_handle(path_or_buf, mode, encoding, compression, memory_map, is_text, errors, storage_options) - 725 codecs.lookup_error(errors) - 727 # open URLs ---> 728 ioargs = _get_filepath_or_buffer( - 729 path_or_buf, - 730 encoding=encoding, - 731 compression=compression, - 732 mode=mode, - 733 storage_options=storage_options, - 734 ) - 736 handle = ioargs.filepath_or_buffer - 737 handles: list[BaseBuffer] - -File /usr/lib/python3/dist-packages/pandas/io/common.py:384, in _get_filepath_or_buffer(filepath_or_buffer, encoding, compression, mode, storage_options) - 382 # assuming storage_options is to be interpreted as headers - 383 req_info = urllib.request.Request(filepath_or_buffer, headers=storage_options) ---> 384 with urlopen(req_info) as req: - 385 content_encoding = req.headers.get("Content-Encoding", None) - 386 if content_encoding == "gzip": - 387 # Override compression based on Content-Encoding header - -File /usr/lib/python3/dist-packages/pandas/io/common.py:289, in urlopen(*args, **kwargs) - 283 """ - 284 Lazy-import wrapper for stdlib urlopen, as that imports a big chunk of - 285 the stdlib. - 286 """ - 287 import urllib.request ---> 289 return urllib.request.urlopen(*args, **kwargs) +Cell In[2], line 18 + 16 except: # recent pandas can read URL without urlopen + 17 url = "http://stats191.stanford.edu/data/salary.table" +---> 18 fh = urlopen(url) + 19 salary_table = pd.read_table(fh) + 20 salary_table.to_csv("salary.table") File /usr/lib/python3.13/urllib/request.py:189, in urlopen(url, data, timeout, context) 187 else: @@ -10063,10 +9924,9 @@ 467 if result is not None: 468 return result -File /usr/lib/python3.13/urllib/request.py:1367, in HTTPSHandler.https_open(self, req) - 1366 def https_open(self, req): --> 1367 return self.do_open(http.client.HTTPSConnection, req, - 1368 context=self._context) +File /usr/lib/python3.13/urllib/request.py:1348, in HTTPHandler.http_open(self, req) + 1347 def http_open(self, req): +-> 1348 return self.do_open(http.client.HTTPConnection, req) File /usr/lib/python3.13/urllib/request.py:1322, in AbstractHTTPHandler.do_open(self, http_class, req, **http_conn_args) 1319 h.request(req.get_method(), req.selector, req.data, headers, @@ -10078,13 +9938,6 @@ URLError: -****************************************************************************** - - -Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/gee_nested_simulation.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/gee_nested_simulation.ipynb - -****************************************************************************** -ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/interactions_anova.ipynb An error occurred while executing the following cell: ------------------ from urllib.request import urlopen @@ -10266,57 +10119,107 @@ URLError: +****************************************************************************** + + +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/glm.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/glm.ipynb +Executing /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/gee_nested_simulation.ipynb to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/examples/notebooks/generated/gee_nested_simulation.ipynb + +****************************************************************************** +ERROR: Error occurred when running /build/reproducible-path/statsmodels-0.14.4+dfsg/examples/notebooks/mstl_decomposition.ipynb An error occurred while executing the following cell: ------------------ -from urllib.request import urlopen -import numpy as np +url = "https://raw.githubusercontent.com/tidyverts/tsibbledata/master/data-raw/vic_elec/VIC2015/demand.csv" +df = pd.read_csv(url) +------------------ -np.set_printoptions(precision=4, suppress=True) -import pandas as pd +--------------------------------------------------------------------------- +ConnectionRefusedError Traceback (most recent call last) +File /usr/lib/python3.13/urllib/request.py:1319, in AbstractHTTPHandler.do_open(self, http_class, req, **http_conn_args) + 1318 try: +-> 1319 h.request(req.get_method(), req.selector, req.data, headers, + 1320 encode_chunked=req.has_header('Transfer-encoding')) + 1321 except OSError as err: # timeout error -pd.set_option("display.width", 100) -import matplotlib.pyplot as plt -from statsmodels.formula.api import ols -from statsmodels.graphics.api import interaction_plot, abline_plot -from statsmodels.stats.anova import anova_lm +File /usr/lib/python3.13/http/client.py:1336, in HTTPConnection.request(self, method, url, body, headers, encode_chunked) + 1335 """Send a complete request to the server.""" +-> 1336 self._send_request(method, url, body, headers, encode_chunked) -try: - salary_table = pd.read_csv("salary.table") -except: # recent pandas can read URL without urlopen - url = "http://stats191.stanford.edu/data/salary.table" - fh = urlopen(url) - salary_table = pd.read_table(fh) - salary_table.to_csv("salary.table") +File /usr/lib/python3.13/http/client.py:1382, in HTTPConnection._send_request(self, method, url, body, headers, encode_chunked) + 1381 body = _encode(body, 'body') +-> 1382 self.endheaders(body, encode_chunked=encode_chunked) -E = salary_table.E -M = salary_table.M -X = salary_table.X -S = salary_table.S ------------------- +File /usr/lib/python3.13/http/client.py:1331, in HTTPConnection.endheaders(self, message_body, encode_chunked) + 1330 raise CannotSendHeader() +-> 1331 self._send_output(message_body, encode_chunked=encode_chunked) +File /usr/lib/python3.13/http/client.py:1091, in HTTPConnection._send_output(self, message_body, encode_chunked) + 1090 del self._buffer[:] +-> 1091 self.send(msg) + 1093 if message_body is not None: + 1094 + 1095 # create a consistent interface to message_body ---------------------------------------------------------------------------- -FileNotFoundError Traceback (most recent call last) -Cell In[2], line 15 - 14 try: ----> 15 salary_table = pd.read_csv("salary.table") - 16 except: # recent pandas can read URL without urlopen +File /usr/lib/python3.13/http/client.py:1035, in HTTPConnection.send(self, data) + 1034 if self.auto_open: +-> 1035 self.connect() + 1036 else: + +File /usr/lib/python3.13/http/client.py:1470, in HTTPSConnection.connect(self) + 1468 "Connect to a host on a given (SSL) port." +-> 1470 super().connect() + 1472 if self._tunnel_host: + +File /usr/lib/python3.13/http/client.py:1001, in HTTPConnection.connect(self) + 1000 sys.audit("http.client.connect", self, self.host, self.port) +-> 1001 self.sock = self._create_connection( + 1002 (self.host,self.port), self.timeout, self.source_address) + 1003 # Might fail in OSs that don't implement TCP_NODELAY + +File /usr/lib/python3.13/socket.py:864, in create_connection(address, timeout, source_address, all_errors) + 863 if not all_errors: +--> 864 raise exceptions[0] + 865 raise ExceptionGroup("create_connection failed", exceptions) + +File /usr/lib/python3.13/socket.py:849, in create_connection(address, timeout, source_address, all_errors) + 848 sock.bind(source_address) +--> 849 sock.connect(sa) + 850 # Break explicitly a reference cycle + +ConnectionRefusedError: [Errno 111] Connection refused + +During handling of the above exception, another exception occurred: + +URLError Traceback (most recent call last) +Cell In[10], line 2 + 1 url = "https://raw.githubusercontent.com/tidyverts/tsibbledata/master/data-raw/vic_elec/VIC2015/demand.csv" +----> 2 df = pd.read_csv(url) File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1026, in read_csv(filepath_or_buffer, sep, delimiter, header, names, index_col, usecols, dtype, engine, converters, true_values, false_values, skipinitialspace, skiprows, skipfooter, nrows, na_values, keep_default_na, na_filter, verbose, skip_blank_lines, parse_dates, infer_datetime_format, keep_date_col, date_parser, date_format, dayfirst, cache_dates, iterator, chunksize, compression, thousands, decimal, lineterminator, quotechar, quoting, doublequote, escapechar, comment, encoding, encoding_errors, dialect, on_bad_lines, delim_whitespace, low_memory, memory_map, float_precision, storage_options, dtype_backend) + 1013 kwds_defaults = _refine_defaults_read( + 1014 dialect, + 1015 delimiter, + (...) + 1022 dtype_backend=dtype_backend, + 1023 ) 1024 kwds.update(kwds_defaults) -> 1026 return _read(filepath_or_buffer, kwds) File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:620, in _read(filepath_or_buffer, kwds) + 617 _validate_names(kwds.get("names", None)) 619 # Create the parser. --> 620 parser = TextFileReader(filepath_or_buffer, **kwds) 622 if chunksize or iterator: + 623 return parser File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1620, in TextFileReader.__init__(self, f, engine, **kwds) + 1617 self.options["has_index_names"] = kwds["has_index_names"] 1619 self.handles: IOHandles | None = None -> 1620 self._engine = self._make_engine(f, self.engine) File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1880, in TextFileReader._make_engine(self, f, engine) + 1878 if "b" not in mode: 1879 mode += "b" -> 1880 self.handles = get_handle( 1881 f, @@ -10329,24 +10232,87 @@ 1888 storage_options=self.options.get("storage_options", None), 1889 ) 1890 assert self.handles is not None + 1891 f = self.handles.handle -File /usr/lib/python3/dist-packages/pandas/io/common.py:873, in get_handle(path_or_buf, mode, encoding, compression, memory_map, is_text, errors, storage_options) - 871 if ioargs.encoding and "b" not in ioargs.mode: - 872 # Encoding ---> 873 handle = open( - 874 handle, - 875 ioargs.mode, - 876 encoding=ioargs.encoding, - 877 errors=errors, - 878 newline="", - 879 ) - 880 else: - 881 # Binary mode +File /usr/lib/python3/dist-packages/pandas/io/common.py:728, in get_handle(path_or_buf, mode, encoding, compression, memory_map, is_text, errors, storage_options) + 725 codecs.lookup_error(errors) + 727 # open URLs +--> 728 ioargs = _get_filepath_or_buffer( + 729 path_or_buf, + 730 encoding=encoding, + 731 compression=compression, + 732 mode=mode, + 733 storage_options=storage_options, + 734 ) + 736 handle = ioargs.filepath_or_buffer + 737 handles: list[BaseBuffer] -FileNotFoundError: [Errno 2] No such file or directory: 'salary.table' +File /usr/lib/python3/dist-packages/pandas/io/common.py:384, in _get_filepath_or_buffer(filepath_or_buffer, encoding, compression, mode, storage_options) + 382 # assuming storage_options is to be interpreted as headers + 383 req_info = urllib.request.Request(filepath_or_buffer, headers=storage_options) +--> 384 with urlopen(req_info) as req: + 385 content_encoding = req.headers.get("Content-Encoding", None) + 386 if content_encoding == "gzip": + 387 # Override compression based on Content-Encoding header -During handling of the above exception, another exception occurred: +File /usr/lib/python3/dist-packages/pandas/io/common.py:289, in urlopen(*args, **kwargs) + 283 """ + 284 Lazy-import wrapper for stdlib urlopen, as that imports a big chunk of + 285 the stdlib. + 286 """ + 287 import urllib.request +--> 289 return urllib.request.urlopen(*args, **kwargs) + +File /usr/lib/python3.13/urllib/request.py:189, in urlopen(url, data, timeout, context) + 187 else: + 188 opener = _opener +--> 189 return opener.open(url, data, timeout) + +File /usr/lib/python3.13/urllib/request.py:489, in OpenerDirector.open(self, fullurl, data, timeout) + 486 req = meth(req) + 488 sys.audit('urllib.Request', req.full_url, req.data, req.headers, req.get_method()) +--> 489 response = self._open(req, data) + 491 # post-process response + 492 meth_name = protocol+"_response" + +File /usr/lib/python3.13/urllib/request.py:506, in OpenerDirector._open(self, req, data) + 503 return result + 505 protocol = req.type +--> 506 result = self._call_chain(self.handle_open, protocol, protocol + + 507 '_open', req) + 508 if result: + 509 return result + +File /usr/lib/python3.13/urllib/request.py:466, in OpenerDirector._call_chain(self, chain, kind, meth_name, *args) + 464 for handler in handlers: + 465 func = getattr(handler, meth_name) +--> 466 result = func(*args) + 467 if result is not None: + 468 return result + +File /usr/lib/python3.13/urllib/request.py:1367, in HTTPSHandler.https_open(self, req) + 1366 def https_open(self, req): +-> 1367 return self.do_open(http.client.HTTPSConnection, req, + 1368 context=self._context) + +File /usr/lib/python3.13/urllib/request.py:1322, in AbstractHTTPHandler.do_open(self, http_class, req, **http_conn_args) + 1319 h.request(req.get_method(), req.selector, req.data, headers, + 1320 encode_chunked=req.has_header('Transfer-encoding')) + 1321 except OSError as err: # timeout error +-> 1322 raise URLError(err) + 1323 r = h.getresponse() + 1324 except: + +URLError: + +An error occurred while executing the following cell: +------------------ +url = "https://raw.githubusercontent.com/tidyverts/tsibbledata/master/data-raw/vic_elec/VIC2015/demand.csv" +df = pd.read_csv(url) +------------------ + +--------------------------------------------------------------------------- ConnectionRefusedError Traceback (most recent call last) File /usr/lib/python3.13/urllib/request.py:1319, in AbstractHTTPHandler.do_open(self, http_class, req, **http_conn_args) 1318 try: @@ -10378,6 +10344,11 @@ -> 1035 self.connect() 1036 else: +File /usr/lib/python3.13/http/client.py:1470, in HTTPSConnection.connect(self) + 1468 "Connect to a host on a given (SSL) port." +-> 1470 super().connect() + 1472 if self._tunnel_host: + File /usr/lib/python3.13/http/client.py:1001, in HTTPConnection.connect(self) 1000 sys.audit("http.client.connect", self, self.host, self.port) -> 1001 self.sock = self._create_connection( @@ -10399,12 +10370,76 @@ During handling of the above exception, another exception occurred: URLError Traceback (most recent call last) -Cell In[2], line 18 - 16 except: # recent pandas can read URL without urlopen - 17 url = "http://stats191.stanford.edu/data/salary.table" ----> 18 fh = urlopen(url) - 19 salary_table = pd.read_table(fh) - 20 salary_table.to_csv("salary.table") +Cell In[10], line 2 + 1 url = "https://raw.githubusercontent.com/tidyverts/tsibbledata/master/data-raw/vic_elec/VIC2015/demand.csv" +----> 2 df = pd.read_csv(url) + +File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1026, in read_csv(filepath_or_buffer, sep, delimiter, header, names, index_col, usecols, dtype, engine, converters, true_values, false_values, skipinitialspace, skiprows, skipfooter, nrows, na_values, keep_default_na, na_filter, verbose, skip_blank_lines, parse_dates, infer_datetime_format, keep_date_col, date_parser, date_format, dayfirst, cache_dates, iterator, chunksize, compression, thousands, decimal, lineterminator, quotechar, quoting, doublequote, escapechar, comment, encoding, encoding_errors, dialect, on_bad_lines, delim_whitespace, low_memory, memory_map, float_precision, storage_options, dtype_backend) + 1013 kwds_defaults = _refine_defaults_read( + 1014 dialect, + 1015 delimiter, + (...) + 1022 dtype_backend=dtype_backend, + 1023 ) + 1024 kwds.update(kwds_defaults) +-> 1026 return _read(filepath_or_buffer, kwds) + +File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:620, in _read(filepath_or_buffer, kwds) + 617 _validate_names(kwds.get("names", None)) + 619 # Create the parser. +--> 620 parser = TextFileReader(filepath_or_buffer, **kwds) + 622 if chunksize or iterator: + 623 return parser + +File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1620, in TextFileReader.__init__(self, f, engine, **kwds) + 1617 self.options["has_index_names"] = kwds["has_index_names"] + 1619 self.handles: IOHandles | None = None +-> 1620 self._engine = self._make_engine(f, self.engine) + +File /usr/lib/python3/dist-packages/pandas/io/parsers/readers.py:1880, in TextFileReader._make_engine(self, f, engine) + 1878 if "b" not in mode: + 1879 mode += "b" +-> 1880 self.handles = get_handle( + 1881 f, + 1882 mode, + 1883 encoding=self.options.get("encoding", None), + 1884 compression=self.options.get("compression", None), + 1885 memory_map=self.options.get("memory_map", False), + 1886 is_text=is_text, + 1887 errors=self.options.get("encoding_errors", "strict"), + 1888 storage_options=self.options.get("storage_options", None), + 1889 ) + 1890 assert self.handles is not None + 1891 f = self.handles.handle + +File /usr/lib/python3/dist-packages/pandas/io/common.py:728, in get_handle(path_or_buf, mode, encoding, compression, memory_map, is_text, errors, storage_options) + 725 codecs.lookup_error(errors) + 727 # open URLs +--> 728 ioargs = _get_filepath_or_buffer( + 729 path_or_buf, + 730 encoding=encoding, + 731 compression=compression, + 732 mode=mode, + 733 storage_options=storage_options, + 734 ) + 736 handle = ioargs.filepath_or_buffer + 737 handles: list[BaseBuffer] + +File /usr/lib/python3/dist-packages/pandas/io/common.py:384, in _get_filepath_or_buffer(filepath_or_buffer, encoding, compression, mode, storage_options) + 382 # assuming storage_options is to be interpreted as headers + 383 req_info = urllib.request.Request(filepath_or_buffer, headers=storage_options) +--> 384 with urlopen(req_info) as req: + 385 content_encoding = req.headers.get("Content-Encoding", None) + 386 if content_encoding == "gzip": + 387 # Override compression based on Content-Encoding header + +File /usr/lib/python3/dist-packages/pandas/io/common.py:289, in urlopen(*args, **kwargs) + 283 """ + 284 Lazy-import wrapper for stdlib urlopen, as that imports a big chunk of + 285 the stdlib. + 286 """ + 287 import urllib.request +--> 289 return urllib.request.urlopen(*args, **kwargs) File /usr/lib/python3.13/urllib/request.py:189, in urlopen(url, data, timeout, context) 187 else: @@ -10433,9 +10468,10 @@ 467 if result is not None: 468 return result -File /usr/lib/python3.13/urllib/request.py:1348, in HTTPHandler.http_open(self, req) - 1347 def http_open(self, req): --> 1348 return self.do_open(http.client.HTTPConnection, req) +File /usr/lib/python3.13/urllib/request.py:1367, in HTTPSHandler.https_open(self, req) + 1366 def https_open(self, req): +-> 1367 return self.do_open(http.client.HTTPSConnection, req, + 1368 context=self._context) File /usr/lib/python3.13/urllib/request.py:1322, in AbstractHTTPHandler.do_open(self, http_class, req, **http_conn_args) 1319 h.request(req.get_method(), req.selector, req.data, headers, @@ -11393,19 +11429,19 @@ [autosummary] generating autosummary for: /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/datasets/statsmodels.datasets.clear_data_home.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/datasets/statsmodels.datasets.get_data_home.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/datasets/statsmodels.datasets.get_rdataset.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/datasets/statsmodels.datasets.webuse.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.__init__.test.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModel.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModelResults.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.LikelihoodModel.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.LikelihoodModelResults.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.Model.rst, ..., /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.var_model.VARResults.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.CointRankResults.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.JohansenTestResult.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECM.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECMResults.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.coint_johansen.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.select_coint_rank.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.select_order.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.x13.x13_arima_analysis.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.x13.x13_arima_select_order.rst [autosummary] generating autosummary for: /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModel.endog_names.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModel.exog_names.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModel.expandparams.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModel.fit.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModel.from_formula.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModel.hessian.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModel.hessian_factor.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModel.information.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModel.initialize.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/dev/generated/statsmodels.base.model.GenericLikelihoodModel.loglike.rst, ..., /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECMResults.test_granger_causality.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECMResults.test_inst_causality.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECMResults.test_normality.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECMResults.test_whiteness.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECMResults.tvalues_alpha.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECMResults.tvalues_beta.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECMResults.tvalues_det_coef.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECMResults.tvalues_det_coef_coint.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECMResults.tvalues_gamma.rst, /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/source/generated/statsmodels.tsa.vector_ar.vecm.VECMResults.var_rep.rst loading intersphinx inventory 'numpy' from https://docs.scipy.org/doc/numpy/objects.inv ... +WARNING: failed to reach any of the inventories with the following issues: +intersphinx inventory 'https://docs.scipy.org/doc/numpy/objects.inv' not fetchable due to : HTTPSConnectionPool(host='docs.scipy.org', port=443): Max retries exceeded with url: /doc/numpy/objects.inv (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) loading intersphinx inventory 'python' from /usr/share/doc/python3-doc/html/objects.inv ... loading intersphinx inventory 'pydagogue' from https://matthew-brett.github.io/pydagogue/objects.inv ... +WARNING: failed to reach any of the inventories with the following issues: +intersphinx inventory 'https://matthew-brett.github.io/pydagogue/objects.inv' not fetchable due to : HTTPSConnectionPool(host='matthew-brett.github.io', port=443): Max retries exceeded with url: /pydagogue/objects.inv (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) loading intersphinx inventory 'matplotlib' from https://matplotlib.org/objects.inv ... loading intersphinx inventory 'scipy' from /usr/share/doc/python-scipy-doc/html/objects.inv ... WARNING: failed to reach any of the inventories with the following issues: intersphinx inventory '/usr/share/doc/python-scipy-doc/html/objects.inv' not fetchable due to : [Errno 2] No such file or directory: '/usr/share/doc/python-scipy-doc/html/objects.inv' -loading intersphinx inventory 'pandas' from /usr/share/doc/python-pandas-doc/html/objects.inv ... -WARNING: failed to reach any of the inventories with the following issues: -intersphinx inventory 'https://matplotlib.org/objects.inv' not fetchable due to : HTTPSConnectionPool(host='matplotlib.org', port=443): Max retries exceeded with url: /objects.inv (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) -WARNING: failed to reach any of the inventories with the following issues: -intersphinx inventory 'https://docs.scipy.org/doc/numpy/objects.inv' not fetchable due to : HTTPSConnectionPool(host='docs.scipy.org', port=443): Max retries exceeded with url: /doc/numpy/objects.inv (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) WARNING: failed to reach any of the inventories with the following issues: -intersphinx inventory 'https://matthew-brett.github.io/pydagogue/objects.inv' not fetchable due to : HTTPSConnectionPool(host='matthew-brett.github.io', port=443): Max retries exceeded with url: /pydagogue/objects.inv (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) +intersphinx inventory 'https://matplotlib.org/objects.inv' not fetchable due to : HTTPSConnectionPool(host='matplotlib.org', port=443): Max retries exceeded with url: /objects.inv (Caused by ProxyError('Unable to connect to proxy', NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))) +loading intersphinx inventory 'pandas' from /usr/share/doc/python-pandas-doc/html/objects.inv ... Writing evaluated template result to /build/reproducible-path/statsmodels-0.14.4+dfsg/docs/build/html/_static/nbsphinx-code-cells.css building [mo]: targets for 0 po files that are out of date writing output... @@ -49093,9 +49129,14 @@ exposure = np.log(exposure) graphics/tests/test_functional.py: 240 warnings - /usr/lib/python3.12/multiprocessing/popen_fork.py:66: DeprecationWarning: This process (pid=1940634) is multi-threaded, use of fork() may lead to deadlocks in the child. + /usr/lib/python3.12/multiprocessing/popen_fork.py:66: DeprecationWarning: This process (pid=205828) is multi-threaded, use of fork() may lead to deadlocks in the child. self.pid = os.fork() +nonparametric/tests/test_kernel_density.py::TestKDEMultivariate::test_continuous_cvls_efficient +nonparametric/tests/test_kernel_density.py::TestKDEMultivariate::test_continuous_cvls_efficient + /usr/lib/python3/dist-packages/joblib/externals/loky/backend/fork_exec.py:38: DeprecationWarning: This process (pid=205828) is multi-threaded, use of fork() may lead to deadlocks in the child. + pid = os.fork() + nonparametric/tests/test_kernel_density.py::TestKDEMultivariateConditional::test_unordered_CV_LS /build/reproducible-path/statsmodels-0.14.4+dfsg/.pybuild/cpython3_3.12_statsmodels/build/statsmodels/nonparametric/kernel_density.py:679: RuntimeWarning: invalid value encountered in scalar divide CV += (G / m_x ** 2) - 2 * (f_X_Y / m_x) @@ -49259,7 +49300,7 @@ W[i], eigva[i], k[i] = util.eigval_decomp(stack_cov[i]) -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html -== 17551 passed, 306 skipped, 139 xfailed, 466 warnings in 1177.19s (0:19:37) == +== 17551 passed, 306 skipped, 139 xfailed, 468 warnings in 3050.41s (0:50:50) == ============================= test session starts ============================== platform linux -- Python 3.13.1, pytest-8.3.4, pluggy-1.5.0 -- /usr/bin/python3.13 cachedir: .pytest_cache @@ -67386,7 +67427,7 @@ exposure = np.log(exposure) graphics/tests/test_functional.py: 240 warnings - /usr/lib/python3.13/multiprocessing/popen_fork.py:67: DeprecationWarning: This process (pid=2143290) is multi-threaded, use of fork() may lead to deadlocks in the child. + /usr/lib/python3.13/multiprocessing/popen_fork.py:67: DeprecationWarning: This process (pid=836211) is multi-threaded, use of fork() may lead to deadlocks in the child. self.pid = os.fork() nonparametric/tests/test_kernel_density.py::TestKDEMultivariateConditional::test_unordered_CV_LS @@ -67552,7 +67593,7 @@ W[i], eigva[i], k[i] = util.eigval_decomp(stack_cov[i]) -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html -== 17551 passed, 306 skipped, 139 xfailed, 466 warnings in 1220.97s (0:20:20) == +== 17551 passed, 306 skipped, 139 xfailed, 466 warnings in 2389.75s (0:39:49) == make[1]: Leaving directory '/build/reproducible-path/statsmodels-0.14.4+dfsg' create-stamp debian/debhelper-build-stamp dh_testroot -O--buildsystem=pybuild @@ -75598,8 +75639,8 @@ make[1]: Leaving directory '/build/reproducible-path/statsmodels-0.14.4+dfsg' dh_sphinxdoc -O--buildsystem=pybuild dh_sphinxdoc: warning: ignoring unknown JavaScript code: debian/python-statsmodels-doc/usr/share/doc/python-statsmodels-doc/html/_static/scripts.js -dh_sphinxdoc: warning: ignoring unknown JavaScript code: debian/python-statsmodels-doc/usr/share/doc/python-statsmodels-doc/html/_static/facebox.js dh_sphinxdoc: warning: ignoring unknown JavaScript code: debian/python-statsmodels-doc/usr/share/doc/python-statsmodels-doc/html/_static/mktree.js +dh_sphinxdoc: warning: ignoring unknown JavaScript code: debian/python-statsmodels-doc/usr/share/doc/python-statsmodels-doc/html/_static/facebox.js dh_installchangelogs -O--buildsystem=pybuild dh_installexamples -O--buildsystem=pybuild dh_installexamples: warning: Cannot auto-detect main package for python-statsmodels-doc. If the default is wrong, please use --doc-main-package @@ -75639,10 +75680,10 @@ dpkg-gencontrol: warning: package python-statsmodels-doc: substitution variable ${sphinxdoc:Built-Using} unused, but is defined dh_md5sums -O--buildsystem=pybuild dh_builddeb -O--buildsystem=pybuild +dpkg-deb: building package 'python3-statsmodels-lib' in '../python3-statsmodels-lib_0.14.4+dfsg-1_arm64.deb'. +dpkg-deb: building package 'python3-statsmodels' in '../python3-statsmodels_0.14.4+dfsg-1_all.deb'. dpkg-deb: building package 'python3-statsmodels-lib-dbgsym' in '../python3-statsmodels-lib-dbgsym_0.14.4+dfsg-1_arm64.deb'. dpkg-deb: building package 'python-statsmodels-doc' in '../python-statsmodels-doc_0.14.4+dfsg-1_all.deb'. -dpkg-deb: building package 'python3-statsmodels' in '../python3-statsmodels_0.14.4+dfsg-1_all.deb'. -dpkg-deb: building package 'python3-statsmodels-lib' in '../python3-statsmodels-lib_0.14.4+dfsg-1_arm64.deb'. dpkg-genbuildinfo --build=binary -O../statsmodels_0.14.4+dfsg-1_arm64.buildinfo dpkg-genchanges --build=binary -O../statsmodels_0.14.4+dfsg-1_arm64.changes dpkg-genchanges: info: binary-only upload (no source code included) @@ -75650,12 +75691,14 @@ dpkg-buildpackage: info: binary-only upload (no source included) dpkg-genchanges: info: including full source code in upload I: copying local configuration +I: user script /srv/workspace/pbuilder/2468496/tmp/hooks/B01_cleanup starting +I: user script /srv/workspace/pbuilder/2468496/tmp/hooks/B01_cleanup finished I: unmounting dev/ptmx filesystem I: unmounting dev/pts filesystem I: unmounting dev/shm filesystem I: unmounting proc filesystem I: unmounting sys filesystem I: cleaning the build env -I: removing directory /srv/workspace/pbuilder/1263307 and its subdirectories -I: Current time: Thu Mar 5 05:58:52 -12 2026 -I: pbuilder-time-stamp: 1772733532 +I: removing directory /srv/workspace/pbuilder/2468496 and its subdirectories +I: Current time: Sat Feb 1 05:28:31 +14 2025 +I: pbuilder-time-stamp: 1738337311