Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
K
km3io
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Package Registry
Container Registry
Model registry
Operate
Environments
Terraform modules
Monitor
Incidents
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
km3py
km3io
Commits
c8f1a399
Commit
c8f1a399
authored
4 years ago
by
Tamas Gal
Browse files
Options
Downloads
Patches
Plain Diff
Cleaned uproot4 transition
parent
7e1cb92c
No related branches found
Branches containing commit
No related tags found
Tags containing commit
1 merge request
!47
Resolve "uproot4 integration"
Pipeline
#16209
failed
4 years ago
Stage: test
Stage: coverage
Stage: doc
Changes
3
Pipelines
1
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
km3io/offline.py
+280
-178
280 additions, 178 deletions
km3io/offline.py
requirements/install.txt
+1
-0
1 addition, 0 deletions
requirements/install.txt
tests/test_offline.py
+83
-59
83 additions, 59 deletions
tests/test_offline.py
with
364 additions
and
237 deletions
km3io/offline.py
+
280
−
178
View file @
c8f1a399
import
binascii
from
collections
import
namedtuple
import
uproot3
import
logging
import
warnings
import
numba
as
nb
import
uproot
import
numpy
as
np
import
awkward
as
ak
from
.definitions
import
mc_header
,
fitparameters
,
reconstruction
from
.definitions
import
mc_header
from
.tools
import
cached_property
,
to_num
,
unfold_indices
from
.rootio
import
Branch
,
BranchMapper
MAIN_TREE_NAME
=
"
E
"
EXCLUDE_KEYS
=
[
"
AAObject
"
,
"
t
"
,
"
fBits
"
,
"
fUniqueID
"
]
# 110 MB based on the size of the largest basket found so far in km3net
BASKET_CACHE_SIZE
=
110
*
1024
**
2
BASKET_CACHE
=
uproot3
.
cache
.
ThreadSafeArrayCache
(
BASKET_CACHE_SIZE
)
def
_nested_mapper
(
key
):
"""
Maps a key in the ROOT file to another key (e.g. trks.pos.x -> pos_x)
"""
return
"
_
"
.
join
(
key
.
split
(
"
.
"
)[
1
:])
EVENTS_MAP
=
BranchMapper
(
name
=
"
events
"
,
key
=
"
Evt
"
,
extra
=
{
"
t_sec
"
:
"
t.fSec
"
,
"
t_ns
"
:
"
t.fNanoSec
"
},
exclude
=
EXCLUDE_KEYS
,
update
=
{
"
n_hits
"
:
"
hits
"
,
"
n_mc_hits
"
:
"
mc_hits
"
,
"
n_tracks
"
:
"
trks
"
,
"
n_mc_tracks
"
:
"
mc_trks
"
,
},
)
SUBBRANCH_MAPS
=
[
BranchMapper
(
name
=
"
tracks
"
,
key
=
"
trks
"
,
extra
=
{},
exclude
=
EXCLUDE_KEYS
+
[
"
trks.usr_data
"
,
"
trks.usr
"
,
"
trks.fUniqueID
"
,
"
trks.fBits
"
],
attrparser
=
_nested_mapper
,
flat
=
False
,
toawkward
=
[
"
fitinf
"
,
"
rec_stages
"
],
),
BranchMapper
(
name
=
"
mc_tracks
"
,
key
=
"
mc_trks
"
,
exclude
=
EXCLUDE_KEYS
+
[
"
mc_trks.rec_stages
"
,
"
mc_trks.fitinf
"
,
"
mc_trks.fUniqueID
"
,
"
mc_trks.fBits
"
,
],
attrparser
=
_nested_mapper
,
toawkward
=
[
"
usr
"
,
"
usr_names
"
],
flat
=
False
,
),
BranchMapper
(
name
=
"
hits
"
,
key
=
"
hits
"
,
exclude
=
EXCLUDE_KEYS
+
[
"
hits.usr
"
,
"
hits.pmt_id
"
,
"
hits.origin
"
,
"
hits.a
"
,
"
hits.pure_a
"
,
"
hits.fUniqueID
"
,
"
hits.fBits
"
,
],
attrparser
=
_nested_mapper
,
flat
=
False
,
),
BranchMapper
(
name
=
"
mc_hits
"
,
key
=
"
mc_hits
"
,
exclude
=
EXCLUDE_KEYS
+
[
"
mc_hits.usr
"
,
"
mc_hits.dom_id
"
,
"
mc_hits.channel_id
"
,
"
mc_hits.tdc
"
,
"
mc_hits.tot
"
,
"
mc_hits.trig
"
,
"
mc_hits.fUniqueID
"
,
"
mc_hits.fBits
"
,
],
attrparser
=
_nested_mapper
,
flat
=
False
,
),
]
class
OfflineBranch
(
Branch
):
@cached_property
def
usr
(
self
):
return
Usr
(
self
.
_mapper
,
self
.
_branch
,
index_chain
=
self
.
_index_chain
)
class
Usr
:
"""
Helper class to access AAObject `usr` stuff (only for events.usr)
"""
log
=
logging
.
getLogger
(
"
offline
"
)
def
__init__
(
self
,
mapper
,
branch
,
index_chain
=
None
):
self
.
_mapper
=
mapper
self
.
_name
=
mapper
.
name
self
.
_index_chain
=
[]
if
index_chain
is
None
else
index_chain
self
.
_branch
=
branch
self
.
_usr_names
=
[]
self
.
_usr_idx_lookup
=
{}
self
.
_usr_key
=
"
usr
"
if
mapper
.
flat
else
mapper
.
key
+
"
.usr
"
self
.
_initialise
()
def
_initialise
(
self
):
try
:
self
.
_branch
[
self
.
_usr_key
]
# This will raise a KeyError in old aanet files
# which has a different strucuter and key (usr_data)
# We do not support those (yet)
except
(
KeyError
,
IndexError
):
print
(
"
The `usr` fields could not be parsed for the
'
{}
'
branch.
"
.
format
(
self
.
_name
)
)
return
self
.
_usr_names
=
[
n
.
decode
(
"
utf-8
"
)
for
n
in
self
.
_branch
[
self
.
_usr_key
+
"
_names
"
].
lazyarray
()[
0
]
]
self
.
_usr_idx_lookup
=
{
name
:
index
for
index
,
name
in
enumerate
(
self
.
_usr_names
)
}
class
OfflineReader
:
"""
reader for offline ROOT files
"""
data
=
self
.
_branch
[
self
.
_usr_key
].
lazyarray
()
event_path
=
"
E/Evt
"
item_name
=
"
OfflineEvent
"
skip_keys
=
[
"
t
"
,
"
AAObject
"
]
aliases
=
{
"
t_sec
"
:
"
t.fSec
"
,
"
t_ns
"
:
"
t.fNanoSec
"
,
"
usr
"
:
"
AAObject/usr
"
,
"
usr_names
"
:
"
AAObject/usr_names
"
,
}
special_branches
=
{
"
hits
"
:
{
"
id
"
:
"
hits.id
"
,
"
channel_id
"
:
"
hits.channel_id
"
,
"
dom_id
"
:
"
hits.dom_id
"
,
"
t
"
:
"
hits.t
"
,
"
tot
"
:
"
hits.tot
"
,
"
trig
"
:
"
hits.trig
"
,
# non-zero if the hit is a triggered hit
},
"
mc_hits
"
:
{
"
id
"
:
"
mc_hits.id
"
,
"
pmt_id
"
:
"
mc_hits.pmt_id
"
,
"
t
"
:
"
mc_hits.t
"
,
# hit time (MC truth)
"
a
"
:
"
mc_hits.a
"
,
# hit amplitude (in p.e.)
"
origin
"
:
"
mc_hits.origin
"
,
# track id of the track that created this hit
"
pure_t
"
:
"
mc_hits.pure_t
"
,
# photon time before pmt simultion
"
pure_a
"
:
"
mc_hits.pure_a
"
,
# amplitude before pmt simution,
"
type
"
:
"
mc_hits.type
"
,
# particle type or parametrisation used for hit
},
"
trks
"
:
{
"
id
"
:
"
trks.id
"
,
"
pos_x
"
:
"
trks.pos.x
"
,
"
pos_y
"
:
"
trks.pos.y
"
,
"
pos_z
"
:
"
trks.pos.z
"
,
"
dir_x
"
:
"
trks.dir.x
"
,
"
dir_y
"
:
"
trks.dir.y
"
,
"
dir_z
"
:
"
trks.dir.z
"
,
"
t
"
:
"
trks.t
"
,
"
E
"
:
"
trks.E
"
,
"
len
"
:
"
trks.len
"
,
"
lik
"
:
"
trks.lik
"
,
"
rec_type
"
:
"
trks.rec_type
"
,
"
rec_stages
"
:
"
trks.rec_stages
"
,
"
fitinf
"
:
"
trks.fitinf
"
,
},
"
mc_trks
"
:
{
"
id
"
:
"
mc_trks.id
"
,
"
pos_x
"
:
"
mc_trks.pos.x
"
,
"
pos_y
"
:
"
mc_trks.pos.y
"
,
"
pos_z
"
:
"
mc_trks.pos.z
"
,
"
dir_x
"
:
"
mc_trks.dir.x
"
,
"
dir_y
"
:
"
mc_trks.dir.y
"
,
"
dir_z
"
:
"
mc_trks.dir.z
"
,
# "status": "mc_trks.status", # TODO: check this
# "mother_id": "mc_trks.mother_id", # TODO: check this
"
type
"
:
"
mc_trks.type
"
,
"
hit_ids
"
:
"
mc_trks.hit_ids
"
,
"
usr
"
:
"
mc_trks.usr
"
,
# TODO: trouble with uproot4
"
usr_names
"
:
"
mc_trks.usr_names
"
,
# TODO: trouble with uproot4
},
}
special_aliases
=
{
"
tracks
"
:
"
trks
"
,
"
mc_tracks
"
:
"
mc_trks
"
,
}
def
__init__
(
self
,
f
,
index_chain
=
None
,
step_size
=
2000
,
keys
=
None
,
aliases
=
None
,
event_ctor
=
None
):
"""
OfflineReader class is an offline ROOT file wrapper
if
self
.
_index_chain
:
data
=
unfold_indices
(
data
,
self
.
_index_chain
)
Parameters
----------
f: str or uproot4.reading.ReadOnlyDirectory (from uproot4.open)
Path to the file of interest or uproot4 filedescriptor.
step_size: int, optional
Number of events to read into the cache when iterating.
Choosing higher numbers may improve the speed but also increases
the memory overhead.
index_chain: list, optional
Keeps track of index chaining.
keys: list or set, optional
Branch keys.
aliases: dict, optional
Branch key aliases.
event_ctor: class or namedtuple, optional
Event constructor.
self
.
_usr_data
=
data
"""
if
isinstance
(
f
,
str
):
self
.
_fobj
=
uproot
.
open
(
f
)
self
.
_filepath
=
f
elif
isinstance
(
f
,
uproot
.
reading
.
ReadOnlyDirectory
):
self
.
_fobj
=
f
self
.
_filepath
=
f
.
_file
.
file_path
else
:
raise
TypeError
(
"
Unsupported file descriptor.
"
)
self
.
_step_size
=
step_size
self
.
_uuid
=
self
.
_fobj
.
_file
.
uuid
self
.
_iterator_index
=
0
self
.
_keys
=
keys
self
.
_event_ctor
=
event_ctor
self
.
_index_chain
=
[]
if
index_chain
is
None
else
index_chain
for
name
in
self
.
_usr_names
:
setattr
(
self
,
name
,
self
[
name
])
# if aliases is not None:
# self.aliases = aliases
# else:
# # Check for usr-awesomeness backward compatibility crap
# if "E/Evt/AAObject/usr" in self._fobj:
# print("Found usr data")
# if ak.count(f["E/Evt/AAObject/usr"].array()) > 0:
# self.aliases.update(
# {
# "usr": "AAObject/usr",
# "usr_names": "AAObject/usr_names",
# }
# )
if
self
.
_keys
is
None
:
self
.
_initialise_keys
()
if
self
.
_event_ctor
is
None
:
self
.
_event_ctor
=
namedtuple
(
self
.
item_name
,
set
(
list
(
self
.
keys
())
+
list
(
self
.
aliases
)
+
list
(
self
.
special_branches
)
+
list
(
self
.
special_aliases
)
),
)
def
__getitem__
(
self
,
item
):
if
self
.
_index_chain
:
return
unfold_indices
(
self
.
_usr_data
,
self
.
_index_chain
)[
:,
self
.
_usr_idx_lookup
[
item
]
]
else
:
return
self
.
_usr_data
[:,
self
.
_usr_idx_lookup
[
item
]]
def
_initialise_keys
(
self
):
skip_keys
=
set
(
self
.
skip_keys
)
toplevel_keys
=
set
(
k
.
split
(
"
/
"
)[
0
]
for
k
in
self
.
_fobj
[
self
.
event_path
].
keys
())
keys
=
(
toplevel_keys
-
skip_keys
).
union
(
list
(
self
.
aliases
.
keys
())
+
list
(
self
.
special_aliases
)
)
for
key
in
list
(
self
.
special_branches
)
+
list
(
self
.
special_aliases
):
keys
.
add
(
"
n_
"
+
key
)
# self._grouped_branches = {k for k in toplevel_keys - skip_keys if isinstance(self._fobj[self.event_path][k].interpretation, uproot.AsGrouped)}
self
.
_keys
=
keys
def
keys
(
self
):
return
self
.
_usr_names
"""
Returns all accessible branch keys, without the skipped ones.
"""
return
self
.
_keys
def
__str__
(
self
):
entries
=
[]
for
name
in
self
.
keys
():
entries
.
append
(
"
{}: {}
"
.
format
(
name
,
self
[
name
]))
return
"
\n
"
.
join
(
entries
)
@property
def
events
(
self
):
# TODO: deprecate this, since `self` is already the container type
return
iter
(
self
)
def
_keyfor
(
self
,
key
):
"""
Return the correct key for a given alias/key
"""
return
self
.
special_aliases
.
get
(
key
,
key
)
def
__getattr__
(
self
,
attr
):
attr
=
self
.
_keyfor
(
attr
)
# if attr in self.keys() or (attr.startswith("n_") and self._keyfor(attr.split("n_")[1]) in self._grouped_branches):
if
attr
in
self
.
keys
():
return
self
.
__getitem__
(
attr
)
raise
AttributeError
(
f
"'
{
self
.
__class__
.
__name__
}
'
object has no attribute
'
{
attr
}
'"
)
def
__repr__
(
self
):
return
"
<{}[{}]>
"
.
format
(
self
.
__class__
.
__name__
,
self
.
_name
)
def
__getitem__
(
self
,
key
):
# indexing
if
isinstance
(
key
,
(
slice
,
int
,
np
.
int32
,
np
.
int64
)):
if
not
isinstance
(
key
,
slice
):
key
=
int
(
key
)
return
self
.
__class__
(
self
.
_fobj
,
index_chain
=
self
.
_index_chain
+
[
key
],
step_size
=
self
.
_step_size
,
aliases
=
self
.
aliases
,
keys
=
self
.
keys
(),
event_ctor
=
self
.
_event_ctor
)
if
isinstance
(
key
,
str
)
and
key
.
startswith
(
"
n_
"
):
# group counts, for e.g. n_events, n_hits etc.
key
=
self
.
_keyfor
(
key
.
split
(
"
n_
"
)[
1
])
arr
=
self
.
_fobj
[
self
.
event_path
][
key
].
array
(
uproot
.
AsDtype
(
"
>i4
"
))
return
unfold_indices
(
arr
,
self
.
_index_chain
)
key
=
self
.
_keyfor
(
key
)
branch
=
self
.
_fobj
[
self
.
event_path
]
# These are special branches which are nested, like hits/trks/mc_trks
# We are explicitly grabbing just a predefined set of subbranches
# and also alias them to be backwards compatible (and attribute-accessible)
if
key
in
self
.
special_branches
:
fields
=
[]
# some fields are not always available, like `usr_names`
for
to_field
,
from_field
in
self
.
special_branches
[
key
].
items
():
if
from_field
in
branch
[
key
].
keys
():
fields
.
append
(
to_field
)
log
.
debug
(
fields
)
out
=
branch
[
key
].
arrays
(
fields
,
aliases
=
self
.
special_branches
[
key
]
)
else
:
out
=
branch
[
self
.
aliases
.
get
(
key
,
key
)].
array
()
class
OfflineReader
:
"""
reader for offline ROOT files
"""
return
unfold_indices
(
out
,
self
.
_index_chain
)
def
__init__
(
self
,
file_path
=
None
):
"""
OfflineReader class is an offline ROOT file wrapper
def
__iter__
(
self
):
self
.
_iterator_index
=
0
self
.
_events
=
self
.
_event_generator
()
return
self
Parameters
----------
file_path : path-like object
Path to the file of interest. It can be a str or any python
path-like object that points to the file.
def
_event_generator
(
self
):
events
=
self
.
_fobj
[
self
.
event_path
]
group_count_keys
=
set
(
k
for
k
in
self
.
keys
()
if
k
.
startswith
(
"
n_
"
))
# special keys to make it easy to count subbranch lengths
log
.
debug
(
"
group_count_keys: %s
"
,
group_count_keys
)
keys
=
set
(
list
(
set
(
self
.
keys
())
-
set
(
self
.
special_branches
.
keys
())
-
set
(
self
.
special_aliases
)
-
group_count_keys
)
+
list
(
self
.
aliases
.
keys
())
)
# all top-level keys for regular branches
log
.
debug
(
"
keys: %s
"
,
keys
)
log
.
debug
(
"
aliases: %s
"
,
self
.
aliases
)
events_it
=
events
.
iterate
(
keys
,
aliases
=
self
.
aliases
,
step_size
=
self
.
_step_size
)
specials
=
[]
special_keys
=
(
self
.
special_branches
.
keys
()
)
# dict-key ordering is an implementation detail
log
.
debug
(
"
special_keys: %s
"
,
special_keys
)
for
key
in
special_keys
:
# print(f"adding {key} with keys {self.special_branches[key].keys()} and aliases {self.special_branches[key]}")
specials
.
append
(
events
[
key
].
iterate
(
self
.
special_branches
[
key
].
keys
(),
aliases
=
self
.
special_branches
[
key
],
step_size
=
self
.
_step_size
,
)
)
group_counts
=
{}
for
key
in
group_count_keys
:
group_counts
[
key
]
=
iter
(
self
[
key
])
log
.
debug
(
"
group_counts: %s
"
,
group_counts
)
for
event_set
,
*
special_sets
in
zip
(
events_it
,
*
specials
):
for
_event
,
*
special_items
in
zip
(
event_set
,
*
special_sets
):
data
=
{}
for
k
in
keys
:
data
[
k
]
=
_event
[
k
]
for
(
k
,
i
)
in
zip
(
special_keys
,
special_items
):
data
[
k
]
=
i
for
tokey
,
fromkey
in
self
.
special_aliases
.
items
():
data
[
tokey
]
=
data
[
fromkey
]
for
key
in
group_counts
:
data
[
key
]
=
next
(
group_counts
[
key
])
yield
self
.
_event_ctor
(
**
data
)
def
__next__
(
self
):
return
next
(
self
.
_events
)
def
__len__
(
self
):
if
not
self
.
_index_chain
:
return
self
.
_fobj
[
self
.
event_path
].
num_entries
elif
isinstance
(
self
.
_index_chain
[
-
1
],
(
int
,
np
.
int32
,
np
.
int64
)):
if
len
(
self
.
_index_chain
)
==
1
:
return
1
# try:
# return len(self[:])
# except IndexError:
# return 1
return
1
else
:
# ignore the usual index magic and access `id` directly
return
len
(
self
.
_fobj
[
self
.
event_path
][
"
id
"
].
array
(),
self
.
_index_chain
)
"""
self
.
_fobj
=
uproot3
.
open
(
file_path
)
self
.
_filename
=
file_path
self
.
_tree
=
self
.
_fobj
[
MAIN_TREE_NAME
]
self
.
_uuid
=
binascii
.
hexlify
(
self
.
_fobj
.
_context
.
uuid
).
decode
(
"
ascii
"
)
def
__actual_len__
(
self
):
"""
The raw number of events without any indexing/slicing magic
"""
return
len
(
self
.
_fobj
[
self
.
event_path
][
"
id
"
].
array
())
def
__repr__
(
self
):
length
=
len
(
self
)
actual_length
=
self
.
__actual_len__
()
return
f
"
{
self
.
__class__
.
__name__
}
(
{
length
}{
'
/
'
+
str
(
actual_length
)
if
length
<
actual_length
else
''
}
events)
"
@property
def
uuid
(
self
):
...
...
@@ -200,21 +312,11 @@ class OfflineReader:
def
__exit__
(
self
,
*
args
):
self
.
close
()
@cached_property
def
events
(
self
):
"""
The `E` branch, containing all offline events.
"""
return
OfflineBranch
(
self
.
_tree
,
mapper
=
EVENTS_MAP
,
subbranchmaps
=
SUBBRANCH_MAPS
)
@cached_property
def
header
(
self
):
"""
The file header
"""
if
"
Head
"
in
self
.
_fobj
:
header
=
{}
for
n
,
x
in
self
.
_fobj
[
"
Head
"
].
_map_3c_string_2c_string_3e_
.
items
():
header
[
n
.
decode
(
"
utf-8
"
)]
=
x
.
decode
(
"
utf-8
"
).
strip
()
return
Header
(
header
)
return
Header
(
self
.
_fobj
[
"
Head
"
].
tojson
()[
"
map<string,string>
"
])
else
:
warnings
.
warn
(
"
Your file header has an unsupported format
"
)
...
...
This diff is collapsed.
Click to expand it.
requirements/install.txt
+
1
−
0
View file @
c8f1a399
...
...
@@ -3,4 +3,5 @@ numba>=0.50
awkward>=1.0.0rc2
awkward0
uproot3>=3.11.1
uproot>=4.0.0rc4
setuptools_scm
This diff is collapsed.
Click to expand it.
tests/test_offline.py
+
83
−
59
View file @
c8f1a399
import
unittest
import
numpy
as
np
from
pathlib
import
Path
import
uuid
import
awkward
as
ak
from
km3net_testdata
import
data_path
from
km3io
import
OfflineReader
from
km3io.offline
import
_nested_mapper
,
Header
from
km3io.offline
import
Header
OFFLINE_FILE
=
OfflineReader
(
data_path
(
"
offline/km3net_offline.root
"
))
OFFLINE_USR
=
OfflineReader
(
data_path
(
"
offline/usr-sample.root
"
))
...
...
@@ -32,7 +34,7 @@ class TestOfflineReader(unittest.TestCase):
assert
self
.
n_events
==
len
(
self
.
r
.
events
)
def
test_uuid
(
self
):
assert
self
.
r
.
uuid
==
"
0001
b192d888fcc711e9b4306cf09e86beef
"
assert
str
(
self
.
r
.
uuid
)
==
"
b192d888
-
fcc7
-
11e9
-
b430
-
6cf09e86beef
"
class
TestHeader
(
unittest
.
TestCase
):
...
...
@@ -147,24 +149,23 @@ class TestOfflineEvents(unittest.TestCase):
def
test_len
(
self
):
assert
self
.
n_events
==
len
(
self
.
events
)
def
test_attributes_available
(
self
):
for
key
in
self
.
events
.
_keymap
.
keys
():
getattr
(
self
.
events
,
key
)
def
test_attributes
(
self
):
assert
self
.
n_events
==
len
(
self
.
events
.
det_id
)
self
.
assertListEqual
(
self
.
det_id
,
list
(
self
.
events
.
det_id
))
print
(
self
.
n_hits
)
print
(
self
.
events
.
hits
)
self
.
assertListEqual
(
self
.
n_hits
,
list
(
self
.
events
.
n_hits
))
self
.
assertListEqual
(
self
.
n_tracks
,
list
(
self
.
events
.
n_tracks
))
self
.
assertListEqual
(
self
.
t_sec
,
list
(
self
.
events
.
t_sec
))
self
.
assertListEqual
(
self
.
t_ns
,
list
(
self
.
events
.
t_ns
))
def
test_keys
(
self
):
assert
np
.
allclose
(
self
.
n_hits
,
self
.
events
[
"
n_hits
"
])
assert
np
.
allclose
(
self
.
n_tracks
,
self
.
events
[
"
n_tracks
"
])
assert
np
.
allclose
(
self
.
t_sec
,
self
.
events
[
"
t_sec
"
])
assert
np
.
allclose
(
self
.
t_ns
,
self
.
events
[
"
t_ns
"
])
assert
np
.
allclose
(
self
.
n_hits
,
self
.
events
[
"
n_hits
"
]
.
tolist
()
)
assert
np
.
allclose
(
self
.
n_tracks
,
self
.
events
[
"
n_tracks
"
]
.
tolist
()
)
assert
np
.
allclose
(
self
.
t_sec
,
self
.
events
[
"
t_sec
"
]
.
tolist
()
)
assert
np
.
allclose
(
self
.
t_ns
,
self
.
events
[
"
t_ns
"
]
.
tolist
()
)
@unittest.skip
def
test_slicing
(
self
):
s
=
slice
(
2
,
8
,
2
)
s_events
=
self
.
events
[
s
]
...
...
@@ -176,20 +177,33 @@ class TestOfflineEvents(unittest.TestCase):
def
test_slicing_consistency
(
self
):
for
s
in
[
slice
(
1
,
3
),
slice
(
2
,
7
,
3
)]:
assert
np
.
allclose
(
self
.
events
[
s
].
n_hits
,
self
.
events
.
n_hits
[
s
])
assert
np
.
allclose
(
self
.
events
[
s
].
n_hits
.
tolist
(),
self
.
events
.
n_hits
[
s
].
tolist
()
)
def
test_index_consistency
(
self
):
for
i
in
[
0
,
2
,
5
]:
assert
np
.
allclose
(
self
.
events
[
i
].
n_hits
,
self
.
events
.
n_hits
[
i
])
assert
np
.
allclose
(
self
.
events
[
i
].
n_hits
,
self
.
events
.
n_hits
[
i
]
)
def
test_index_chaining
(
self
):
assert
np
.
allclose
(
self
.
events
[
3
:
5
].
n_hits
,
self
.
events
.
n_hits
[
3
:
5
])
assert
np
.
allclose
(
self
.
events
[
3
:
5
][
0
].
n_hits
,
self
.
events
.
n_hits
[
3
:
5
][
0
])
assert
np
.
allclose
(
self
.
events
[
3
:
5
].
hits
[
1
].
dom_id
[
4
]
,
self
.
events
.
hits
[
3
:
5
]
[
1
][
4
].
dom_id
self
.
events
[
3
:
5
].
n_
hits
.
tolist
()
,
self
.
events
.
n_
hits
[
3
:
5
]
.
tolist
()
)
assert
np
.
allclose
(
self
.
events
.
hits
[
3
:
5
][
1
][
4
].
dom_id
,
self
.
events
[
3
:
5
][
1
][
4
].
hits
.
dom_id
self
.
events
[
3
:
5
][
0
].
n_hits
,
self
.
events
.
n_hits
[
3
:
5
][
0
]
)
@unittest.skip
def
test_index_chaining_on_nested_branches_aka_records
(
self
):
assert
np
.
allclose
(
self
.
events
[
3
:
5
].
hits
[
1
].
dom_id
[
4
],
self
.
events
.
hits
[
3
:
5
][
1
][
4
].
dom_id
,
)
assert
np
.
allclose
(
self
.
events
.
hits
[
3
:
5
][
1
][
4
].
dom_id
.
tolist
(),
self
.
events
[
3
:
5
][
1
][
4
].
hits
.
dom_id
.
tolist
(),
)
def
test_fancy_indexing
(
self
):
...
...
@@ -200,15 +214,17 @@ class TestOfflineEvents(unittest.TestCase):
assert
8
==
len
(
first_tracks
.
rec_stages
)
assert
8
==
len
(
first_tracks
.
lik
)
@unittest.skip
def
test_iteration
(
self
):
i
=
0
for
event
in
self
.
events
:
i
+=
1
assert
10
==
i
@unittest.skip
def
test_iteration_2
(
self
):
n_hits
=
[
e
.
n_hits
for
e
in
self
.
events
]
assert
np
.
allclose
(
n_hits
,
self
.
events
.
n_
hits
)
n_hits
=
[
len
(
e
.
hits
.
id
)
for
e
in
self
.
events
]
assert
np
.
allclose
(
n_hits
,
ak
.
num
(
self
.
events
.
hits
.
id
,
axis
=
1
).
tolist
()
)
def
test_str
(
self
):
assert
str
(
self
.
n_events
)
in
str
(
self
.
events
)
...
...
@@ -274,16 +290,14 @@ class TestOfflineHits(unittest.TestCase):
],
}
def
test_attributes
_available
(
self
):
for
key
in
self
.
hits
.
_keymap
.
keys
()
:
def
test_
fields_work_as_keys_and_
attributes
(
self
):
for
key
in
self
.
hits
.
fields
:
getattr
(
self
.
hits
,
key
)
self
.
hits
[
key
]
def
test_channel_ids
(
self
):
self
.
assertTrue
(
all
(
c
>=
0
for
c
in
self
.
hits
.
channel_id
.
min
()))
self
.
assertTrue
(
all
(
c
<
31
for
c
in
self
.
hits
.
channel_id
.
max
()))
def
test_str
(
self
):
assert
str
(
self
.
n_hits
)
in
str
(
self
.
hits
)
self
.
assertTrue
(
all
(
c
>=
0
for
c
in
ak
.
min
(
self
.
hits
.
channel_id
,
axis
=
1
)))
self
.
assertTrue
(
all
(
c
<
31
for
c
in
ak
.
max
(
self
.
hits
.
channel_id
,
axis
=
1
)))
def
test_repr
(
self
):
assert
str
(
self
.
n_hits
)
in
repr
(
self
.
hits
)
...
...
@@ -292,7 +306,7 @@ class TestOfflineHits(unittest.TestCase):
for
idx
,
dom_id
in
self
.
dom_id
.
items
():
self
.
assertListEqual
(
dom_id
,
list
(
self
.
hits
.
dom_id
[
idx
][:
len
(
dom_id
)]))
for
idx
,
t
in
self
.
t
.
items
():
assert
np
.
allclose
(
t
,
self
.
hits
.
t
[
idx
][:
len
(
t
)])
assert
np
.
allclose
(
t
,
self
.
hits
.
t
[
idx
][:
len
(
t
)]
.
tolist
()
)
def
test_slicing
(
self
):
s
=
slice
(
2
,
8
,
2
)
...
...
@@ -306,28 +320,39 @@ class TestOfflineHits(unittest.TestCase):
def
test_slicing_consistency
(
self
):
for
s
in
[
slice
(
1
,
3
),
slice
(
2
,
7
,
3
)]:
for
idx
in
range
(
3
):
assert
np
.
allclose
(
self
.
hits
.
dom_id
[
idx
][
s
],
self
.
hits
[
idx
].
dom_id
[
s
])
assert
np
.
allclose
(
OFFLINE_FILE
.
events
[
idx
].
hits
.
dom_id
[
s
],
self
.
hits
.
dom_id
[
idx
][
s
]
self
.
hits
.
dom_id
[
idx
][
s
].
tolist
(),
self
.
hits
[
idx
].
dom_id
[
s
].
tolist
()
)
assert
np
.
allclose
(
OFFLINE_FILE
.
events
[
idx
].
hits
.
dom_id
[
s
].
tolist
(),
self
.
hits
.
dom_id
[
idx
][
s
].
tolist
(),
)
def
test_index_consistency
(
self
):
for
idx
,
dom_ids
in
self
.
dom_id
.
items
():
assert
np
.
allclose
(
self
.
hits
[
idx
].
dom_id
[:
self
.
n_hits
],
dom_ids
[:
self
.
n_hits
]
self
.
hits
[
idx
].
dom_id
[:
self
.
n_hits
]
.
tolist
()
,
dom_ids
[:
self
.
n_hits
]
)
assert
np
.
allclose
(
OFFLINE_FILE
.
events
[
idx
].
hits
.
dom_id
[:
self
.
n_hits
],
OFFLINE_FILE
.
events
[
idx
].
hits
.
dom_id
[:
self
.
n_hits
]
.
tolist
()
,
dom_ids
[:
self
.
n_hits
],
)
for
idx
,
ts
in
self
.
t
.
items
():
assert
np
.
allclose
(
self
.
hits
[
idx
].
t
[:
self
.
n_hits
],
ts
[:
self
.
n_hits
])
assert
np
.
allclose
(
OFFLINE_FILE
.
events
[
idx
].
hits
.
t
[:
self
.
n_hits
],
ts
[:
self
.
n_hits
]
self
.
hits
[
idx
].
t
[:
self
.
n_hits
].
tolist
(),
ts
[:
self
.
n_hits
]
)
assert
np
.
allclose
(
OFFLINE_FILE
.
events
[
idx
].
hits
.
t
[:
self
.
n_hits
].
tolist
(),
ts
[:
self
.
n_hits
],
)
def
test_keys
(
self
):
assert
"
dom_id
"
in
self
.
hits
.
keys
()
def
test_fields
(
self
):
assert
"
dom_id
"
in
self
.
hits
.
fields
assert
"
channel_id
"
in
self
.
hits
.
fields
assert
"
t
"
in
self
.
hits
.
fields
assert
"
tot
"
in
self
.
hits
.
fields
assert
"
trig
"
in
self
.
hits
.
fields
assert
"
id
"
in
self
.
hits
.
fields
class
TestOfflineTracks
(
unittest
.
TestCase
):
...
...
@@ -337,16 +362,9 @@ class TestOfflineTracks(unittest.TestCase):
self
.
tracks_numucc
=
OFFLINE_NUMUCC
self
.
n_events
=
10
def
test_attributes_available
(
self
):
for
key
in
self
.
tracks
.
_keymap
.
keys
():
getattr
(
self
.
tracks
,
key
)
@unittest.skip
def
test_attributes
(
self
):
for
idx
,
dom_id
in
self
.
dom_id
.
items
():
self
.
assertListEqual
(
dom_id
,
list
(
self
.
hits
.
dom_id
[
idx
][:
len
(
dom_id
)]))
for
idx
,
t
in
self
.
t
.
items
():
assert
np
.
allclose
(
t
,
self
.
hits
.
t
[
idx
][:
len
(
t
)])
def
test_fields
(
self
):
for
field
in
[
'
id
'
,
'
pos_x
'
,
'
pos_y
'
,
'
pos_z
'
,
'
dir_x
'
,
'
dir_y
'
,
'
dir_z
'
,
'
t
'
,
'
E
'
,
'
len
'
,
'
lik
'
,
'
rec_type
'
,
'
rec_stages
'
,
'
fitinf
'
]:
getattr
(
self
.
tracks
,
field
)
def
test_item_selection
(
self
):
self
.
assertListEqual
(
...
...
@@ -354,8 +372,9 @@ class TestOfflineTracks(unittest.TestCase):
)
def
test_repr
(
self
):
assert
"
10
"
in
repr
(
self
.
tracks
)
assert
"
10
*
"
in
repr
(
self
.
tracks
)
@unittest.skip
def
test_slicing
(
self
):
tracks
=
self
.
tracks
self
.
assertEqual
(
10
,
len
(
tracks
))
# 10 events
...
...
@@ -375,6 +394,7 @@ class TestOfflineTracks(unittest.TestCase):
list
(
tracks
.
E
[:,
0
][
_slice
]),
list
(
tracks
[
_slice
].
E
[:,
0
])
)
@unittest.skip
def
test_nested_indexing
(
self
):
self
.
assertAlmostEqual
(
self
.
f
.
events
.
tracks
.
fitinf
[
3
:
5
][
1
][
9
][
2
],
...
...
@@ -398,15 +418,18 @@ class TestBranchIndexingMagic(unittest.TestCase):
def
setUp
(
self
):
self
.
events
=
OFFLINE_FILE
.
events
def
test_
foo
(
self
):
def
test_
slicing_magic
(
self
):
self
.
assertEqual
(
318
,
self
.
events
[
2
:
4
].
n_hits
[
0
])
assert
np
.
allclose
(
self
.
events
[
3
].
tracks
.
dir_z
[
10
],
self
.
events
.
tracks
.
dir_z
[
3
,
10
]
)
assert
np
.
allclose
(
self
.
events
[
3
:
6
].
tracks
.
pos_y
[:,
0
],
self
.
events
.
tracks
.
pos_y
[
3
:
6
,
0
]
self
.
events
[
3
:
6
].
tracks
.
pos_y
[:,
0
].
tolist
(),
self
.
events
.
tracks
.
pos_y
[
3
:
6
,
0
].
tolist
(),
)
@unittest.skip
def
test_selecting_specific_items_via_a_list
(
self
):
# test selecting with a list
self
.
assertEqual
(
3
,
len
(
self
.
events
[[
0
,
2
,
3
]]))
...
...
@@ -415,9 +438,11 @@ class TestUsr(unittest.TestCase):
def
setUp
(
self
):
self
.
f
=
OFFLINE_USR
@unittest.skip
def
test_str_flat
(
self
):
print
(
self
.
f
.
events
.
usr
)
@unittest.skip
def
test_keys_flat
(
self
):
self
.
assertListEqual
(
[
...
...
@@ -439,27 +464,29 @@ class TestUsr(unittest.TestCase):
"
NGeometryVetoHits
"
,
"
ClassficationScore
"
,
],
self
.
f
.
events
.
usr
.
keys
(),
self
.
f
.
events
.
usr
.
keys
()
.
tolist
()
,
)
@unittest.skip
def
test_getitem_flat
(
self
):
assert
np
.
allclose
(
[
118.6302815337638
,
44.33580521344907
,
99.93916717621543
],
self
.
f
.
events
.
usr
[
"
CoC
"
],
self
.
f
.
events
.
usr
[
"
CoC
"
]
.
tolist
()
,
)
assert
np
.
allclose
(
[
37.51967774166617
,
-
10.280346193553832
,
13.67595659707355
],
self
.
f
.
events
.
usr
[
"
DeltaPosZ
"
],
self
.
f
.
events
.
usr
[
"
DeltaPosZ
"
]
.
tolist
()
,
)
@unittest.skip
def
test_attributes_flat
(
self
):
assert
np
.
allclose
(
[
118.6302815337638
,
44.33580521344907
,
99.93916717621543
],
self
.
f
.
events
.
usr
.
CoC
,
self
.
f
.
events
.
usr
.
CoC
.
tolist
()
,
)
assert
np
.
allclose
(
[
37.51967774166617
,
-
10.280346193553832
,
13.67595659707355
],
self
.
f
.
events
.
usr
.
DeltaPosZ
,
self
.
f
.
events
.
usr
.
DeltaPosZ
.
tolist
()
,
)
...
...
@@ -467,18 +494,20 @@ class TestMcTrackUsr(unittest.TestCase):
def
setUp
(
self
):
self
.
f
=
OFFLINE_MC_TRACK_USR
@unittest.skip
def
test_usr_names
(
self
):
n_tracks
=
len
(
self
.
f
.
events
)
for
i
in
range
(
3
):
self
.
assertListEqual
(
[
b
"
bx
"
,
b
"
by
"
,
b
"
ichan
"
,
b
"
cc
"
],
[
"
bx
"
,
"
by
"
,
"
ichan
"
,
"
cc
"
],
self
.
f
.
events
.
mc_tracks
.
usr_names
[
i
][
0
].
tolist
(),
)
self
.
assertListEqual
(
[
b
"
energy_lost_in_can
"
],
[
"
energy_lost_in_can
"
],
self
.
f
.
events
.
mc_tracks
.
usr_names
[
i
][
1
].
tolist
(),
)
@unittest.skip
def
test_usr
(
self
):
assert
np
.
allclose
(
[
0.0487
,
0.0588
,
3
,
2
],
...
...
@@ -488,8 +517,3 @@ class TestMcTrackUsr(unittest.TestCase):
assert
np
.
allclose
(
[
0.147
,
0.4
,
3
,
2
],
self
.
f
.
events
.
mc_tracks
.
usr
[
1
][
0
].
tolist
(),
atol
=
0.001
)
class
TestNestedMapper
(
unittest
.
TestCase
):
def
test_nested_mapper
(
self
):
self
.
assertEqual
(
"
pos_x
"
,
_nested_mapper
(
"
trks.pos.x
"
))
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment