-
Notifications
You must be signed in to change notification settings - Fork 7
arokem/vset
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
The Visual System Engineering Toolbox (VSET) is a Matlab toolbox designed for calculating the properties of the front end of the visual system. The VSET code is a portion of Image Systems Engineering Toolbox (ISET) that is sold by Imageval Consulting, LLC. That code is designed to help industrial partners design novel image sensors. The VSET portion of the ISET code is freely distributed for use in modeling image formation in biological systems. For a general introduction to human vision, please see: http://foundationsofvision.stanford.edu/ For examples of ISET/VSET code and tutorials on human vision, please see: https://www.stanford.edu/group/vista/cgi-bin/FOV/computational-examples/ Copyright (c) 2012, Brian A. Wandell wandell@stanford.edu Implementation notes The toolbox is written around several data structures. Each is implemented using set/get/create syntax. Major computations are implemented by computing with these data structures. The scene data structure describes the scene radiance (photons). It is set up to permit depth encoding, though nearly all of the current examples are based on a radiance field originating from a single plane. The optical image transforms the irradiance distribution at the sensor, after the radiance has passed through the topics. There are several computational models that implement the transformation: diffraction limited, shift invariant, and ray trace (shift variant). You can choose the one you want to use depending on the level of information you have about the optics. There is an implementation of the optics of the human eye, based on work from Marimont and Wandell. This toolbox is coordinated with the Wavefront Toolbox (also in github) uses data from adaptive optics to simulate the defocus based on measurements of many different human eyes. The sensor transforms the irradiance into a spatial array of cone absorptions. The sensor pixels and spectral quantum efficiency can be set to model the human cones at various eccentricities and with various types of inert pigments (macular pigment, lens density, optical density). The sensor simulates the photon absorptions in the cone (and rod) mosaics. *** END README
About
visual system engineering tools
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published