From 59ab31e6cba524f8305086bfee67442726b62e93 Mon Sep 17 00:00:00 2001 From: Oliver Ruebel Date: Tue, 20 Aug 2024 11:19:18 -0700 Subject: [PATCH] Update docs/gallery/advanced_io/plot_linking_data.py Co-authored-by: Steph Prince <40640337+stephprince@users.noreply.github.com> --- docs/gallery/advanced_io/plot_linking_data.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/gallery/advanced_io/plot_linking_data.py b/docs/gallery/advanced_io/plot_linking_data.py index b90e60f84..9f3030a2f 100644 --- a/docs/gallery/advanced_io/plot_linking_data.py +++ b/docs/gallery/advanced_io/plot_linking_data.py @@ -271,8 +271,8 @@ # the file stystem does not allow for large files. While we can # achieve this by writing different components (e.g., :py:meth:`~pynwb.base.TimeSeries`) to different files as described above, # this option does not allow splitting data from single datasets. An alternative option is to use the -# `family` driver in `h5py` to automatically split the NWB file into a collection of many HDF5 files. -# The `family` driver stores the file on disk as a series of fixed-length chunks (each in its own file). +# ``family`` driver in ``h5py`` to automatically split the NWB file into a collection of many HDF5 files. +# The ``family`` driver stores the file on disk as a series of fixed-length chunks (each in its own file). # In practice, to write very large arrays, we can combine this approach with :ref:`iterative_write` to # avoid having to load all data into memory. In the example shown here we use a manual approach to # iterative write by using :py:class:`~hdmf.backends.hdf5.h5_utils.H5DataIO` to create an empty dataset and