Abstract

Since May 2006, the two STELLA robotic telescopes at the Izaña observatory in Tenerife, Spain, delivered an almost uninterrupted stream of scientific data. To achieve such a high level of autonomous operation, the replacement of all troubleshooting skills of a regular observer in software was required. Care must be taken on error handling issues and on robustness of the algorithms used. In the current paper, we summarize the approaches we followed in the STELLA observatory.

1. Introduction

STELLA is a fully autonomous robotic observatory with two 1.2 m az-alt telescopes located at the Izaña observatory in Tenerife, Spain [1, 2]. STELLA-I is a classical f/8 Cassegrain-type telescope, equipped with a swiveling M3 mirror to make both Nasmith foci available. STELLA-II is a highly specialized telescope with the single purpose to feed as much light as possibleinto an on-axis fiber. It has a single spherical mirror at f/1.95. At F1, a field corrector matches the PSF to the fiber entrance f-ratio and additionally provides a field of view of roughly 2 arcmin around the fiber. For acquisition and guiding, STELLA-II features an auxiliary, 15 cm refracting telescope. STELLA-I currently feeds the STELLA Echelle spectrograph (SES, see [3]), but in late 2009 the new wide-field imaging photometer (WiFSIP) should be operated on this telescope. STELLA-II is currently in a testing phase.

Both telescopes are truly robotic telescopes in the sense that they autonomously react to changing weather conditions, including operation of the telescope sheltering building. The object selection is not based on a single, prescribed sequence, but is always constructed dynamically, making fast reaction easy (see Section 2). Internet connection is only necessary for data retrieval and upload of new targets to the target pool, both of which can be achieved with considerable low bandwidth. Remote observing is possible but has so far never been necessary for normal scientific observations.

In the next sections, a step-by-step receipt of how a human observer has been replaced by individual pieces of software is presented.

2. Scheduling Observations

Different to classic observatories, robotic observatories in general and STELLA especially, do not divide the available observing time into different time slices and dedicate these to single users. All targets are active at any time, allowing for observing campaigns that span months and, quite usually, even years. An overview on the different projects conducted at STELLA might help to mitigate the scheduling requirements. One of the key projects on the spectroscopic telescope is the investigation of stellar magnetic activity on a handful of stars. The rotational period defines the time scale at which a couple of, say, twenty observations should occur, at intervals as regular as possible. The usability of a single observation highly depends on all the other observations occurring around it. Only if a high phase coverage could be achieved, the individual observation was useful. Consequentially, the scheduling algorithm must be able to adjust the priority of such an observing project. Once started, it should be assured that it can also finish in time. If in doubt, refrain from starting the project at all. What comes to ease here is the general insensitiveness to the starting point of such programs, and the only focus lies on a proper phase coverage.

Studying highly phase critical phenomena like the Blazhko effect in RR Lyrae star requires the observations to be timed exactly around certain, well-known phases. Here, no freedom in choosing the starting point is possible, but observations do not need to be clumped especially close together.

Observations of radial velocity curves of Cephei stars or extrasolar planets relax that even further. Here you aim at good phase coverage, but the spacing between individual observations does hardly matter.

Objects with a prior unknown periods should be scheduled such that no bias is introduced on subsequent period determination.

The final class of targets is those that introduce no special timing or periodicity but rather yield the highest importance if observed at optimal conditions—optimal can refer to certain astronomic conditions, like no moon-light pollution or minimum airmass (easily to predict) or certain seeing requirements (difficult to predict).

Algorithms that deal with optimized scheduling are not confined to robotic telescopes alone. Basically all robotic processes face similar problems (e.g., [4]). Very different approaches to the scheduling problem exist in the literature; those relevant to robotic telescopes will be discussed briefly.

2.1. Queue Scheduling

Queue scheduling is the simplest approach possible, but also the least flexible. A superior process, most likely a human, defines a schedule for the upcoming observing period (not too long to make reaction to bad weather periods possible, not too short to gain advantage of the automated observing process). On observation start, the queue is loaded into the system and followed task by task. A coupling with additional constraints (target must be above a certain height; target may only be observed within an hour of the scheduled time) may allow for a limited degree of flexibility, which make queue scheduling apt for single-task surveillance projects.

2.2. Critical-Path Scheduling

A scheduling algorithm that splits a single task, like an entire observing campaign, into different, atomic, subtasks with various dependencies amongst each other is known as critical-path scheduling (e.g., [5]). It can be seen as the mathematical description of a Gantt chart and is mainly used in huge construction projects, where the main focus lies on the dependency between individual subtasks. Still, the main task of scheduling the different task relative to each other remains and such critical-path scheduling is seldom used in robotic astronomy.

2.3. Optimal Scheduling

An optimal schedule describes the particular flow of observations that allows the maximization of the scientific return measured according to a predefined metrics like shutter-open time. The high number of possible permutations of targets, , makes any algorithm a highly demanding one. In particular, unpredictable changes in environmental conditions break the optimal schedule, and a recalculation is necessary. At ground-based observatories, an optimal scheduling schema is difficult to implement due to unforeseen changes in weather conditions. The Hubble Space Telescope, however, uses a software package called SPIKE [6] that delivers an optimal schedule for 14-day periods. Attempts to use the same algorithm on ground-based facilities, for example, with the Very Large Telescope [7] or the Subaru telescope [8] have yielded some success in producing guidelines for night astronomers.

2.4. Dispatch Scheduling

The algorithm that schedules targets in real time, always according to the current (observational) conditions, is known as dispatch scheduling. From the entire pool of targets available, the algorithm calculates a per-target merit, picking then the target with the highest yield.

The summation is done over individual merits , with different weights for a particular merit. A balanced choice on the individual weights and merits allows for a very capable scheduling algorithm. This approach allows easy reaction on changing weather conditions and at the same time optimizes target scheduling for distinct side goals. However, it has no predictive capabilities in the sense that the currently top-rated observation will be done, regardless of an even higher yield possible in the future. Nevertheless, this approach is probably suited best for ground-based telescopes as the reaction to changing weather conditions is algorithm-inherent. Dispatch scheduling is thus used in many robotic telescope (e.g., [9]). On the STELLA observatory it is applied in a somewhat modified approach to compensate for the bad long-time behavior, see.

Here, the 's, weighted with constant factors , allow long-term modification of target selection (i.e., over several nights), while the 's are mainly used for short-term scheduling, that is, over the course of a given night. On STELLA, the target itself defines which merits it may use. This is similar to setting all weights on all nonspecified merits to zero but allows easier adaptation to new observing strategies: new merits may be added at any time, given that they are available at run time. Out of convenience, all 's and 's are limited within , but merits exceeding one are allowed by adjustment of the weights. Generally, three sets of weights for all 's and 's are available, reflecting the three principal priority levels: level A for high-priority targets; all observations requested for a single target must be completed to allow scientific conclusions. Level B indicates mid-priority targets. Observing strategy is best-effort based, with (currently) a high likeliness of all observations to be completed. The lowest priority, , is designed for targets that either add some scientific value, if observed, or for low-priority targets in large surveys. As a matter of fact, targets within this priority class are mostly observed during partly clouded nights, when targets in higher priority bins fail.

From the vast possibilities that the different combinations of weights and merits allow, only a few possible setups are currently implemented on STELLA. Large surveys on different targets have a pure airmass merit that peaks at one on the target's culmination and drops to zero at the apparent horizon, which lies between and . Depending on the priority class, the weights are (A, B, C). Once observed successfully, their toggle-merit drops to zero. Surveys, where targets should be revisited after a certain, maybe target-depending, period have an that starts at one, drops to zero on a successful observation, and linearly regains its top value of one after the specified period.

Targets that need observation strictly at specific phases utilize a so-called phase merit, that is the sum of normalized Gaussian centered at the requested phases. Targets that provide no phase zero point are scheduled according to their first successful observation. Again the weights read as (A,B,C). In the basic form, where only strict phasing, but not the evolved time span matters, the single is constructed such that it starts at , where is the number of phases requested. As the number of phases successfully observed increases, follows . The accompanying weight equals , leading to a gentle increase in total merit from one to once the target has been started.

The most ambitious scheduling is done for objects that need a couple of phases observed within a few periods. At STELLA Doppler-imaging targets, these periods are in the range of several days, leaving you 2-3 weeks to complete a target. The merit is constructed in the following way. We start with an that peaks around the opposition of the target, mathematically a normalized Gaussian with an FWHM of three periods. Once such a target is successfully observed for the first time, a second gets activated, which is a parabola fit through points , the reciprocal of the weight at the starting time and through zero after some multiple of the period, default after . Additionally, a time until the maximum is reached can be specified. In addition with the beforementioned phase merit, a complex merit functions like the one depicted in Figure 1 are reached. Some more details can be found in [10].

3. How to Carry out a Robotic Observation

When a target has been selected by the scheduling process, all the individual subtasks for that particular target have to be carried out—in parallel whenever possible, strictly sequential if dependencies between the subtasks exist. Parallelizing many tasks can save valuable observing time, on STELLA, read-out of the scientific CCD takes place while the telescope already slews to the next target. In pretty exactly half of the cases, the next target is already acquired and closed-loop guiding has commenced, when the read-out finishes.

STELLA houses a spectroscopic and an imaging telescope, thus the individual subtasks for a single observation differ quite substantially, nevertheless the general idea of splitting an observation into subtasks with the possibility to execute them in parallel or sequentially stays the same. In STELLA, we implemented a generic sequencing schema, which is described in detail in [11]. There we also explain how targets can define their own observing sequence and how a template sequence is constructed. In this paper, however, we want to focus more on the individual subtasks and their implementation, particularly pointing out that the solutions have been chosen for robustness rather than high accuracy. In the description, we follow the principle time line of an astronomic observation, starting by judging the overall weather situation, followed by pointing and focusing of the telescope, then acquiring of a target, followed by closed-loop guiding during the scientific exposure. These sections apply particularly to the fiber-fed instrument Stella-I. Tasks only required for the imaging telescopes Stella-II follow in the next Section 4.

3.1. Protecting the Telescope in Harsh Weather Conditions

Protecting the telescope from the outside during bad weather conditions is one of the task that an autonomous observatory has to fulfill with the highest possible reliability. Needless to say that it is better to lose some observing time at unstable weather conditions than to risk damage to the telescope or the instrumentation due to high humidity, rain, high wind, or, particularly cumbersome in Tenerife, dust. The STELLA observatory is laid out as a building with a roll-off roof with two, elliptically shaped roof halves driven by crane motors. Opening and closing of the roof is possible in any position of the two telescopes; see Figure 2. As an additional security mechanism, a simple watchdog system that automatically closes the roof in case of computer crashes is used. Systems where the telescope has to be moved to a certain position before the closing of the roof can commence should be avoided whenever possible, as it leaves the instruments in an unprotected state if the telescope, for whatever reason, cannot be turned.

In standard, that is, unattended operation mode, the meteorological readings of two independent weather stations decide on the opening or closing of the roof. All sensors considered critical (precipitation, humidity, temperature, and wind speed) are available at either station, with the redundancy allowing for operation of the observatory even in case of failure of one of the stations. The building control acts completely independent from all other systems and had been installed even before the telescopes was put in place. In almost eight years of building operation, it never failed to protect the telescope.

The decision-making process for the open/close roof process relies on the current reading of the four critical sensors, as well as on the history of their measures. A solar-height sensor is a purely calculated sensor but enters just like another critical sensor. Combining two or more sensors in the decision making turned out to be not necessary.

Most of the critical sensors act as Schmidt-triggers; they toggle their weather state on two values, depending on the general direction of the past measures. Toggling from one state into the other is additionally delayed by a sensor-dependant retard time, during that the sensor's reading must stay in the new state, otherwise toggling does not occur. Typically, this retard time is two seconds or four read cycles if toggling to the bad weather state—its main purpose is to filter out erroneous readings of a sensor. Toggling to the good weather state is more depending on the site characteristics. For Izaña, humidity toggles to bad at 70% the clear toggle is set at 60% Wind speed toggles at 20 m/s and 10 m/s, respectively, temperature at −2°C and +1°C. Precipitation has just an on/off state, thus zero is considered no rain, one is considered raining. A brightness sensor toggles at 3000 lx and 10000 lx. It is mainly a guard against wrong-posed solar height calculations.

We apply a retard time of two hours for a humidity event, four hours for a rain event, and twenty minutes for a high wind-gust event. This principle is depicted in Figure 3(a), for reaction on the humidity.

Early on, the exposed location of the Tenerife site, just above the inversion layer, demanded for the capabilities to predict inversion layer breakdowns. For that, the course of the humidity is examined in more detail and a prognosis for the next 100 seconds—the closing time of the roofs—is derived from it; see Figure 3(b). Different time bases and either linear or parabolic extrapolation are used for the near-future humidity development. Only if at least six out of seven extrapolators predict a humidity above 80%—compared to the normal toggle of 70%—, the roofs close.

3.2. Pointing the Telescope

For a reliable object acquisition, a good initial pointing of the telescope is desirable. For that, a stable mount is inevitable. To compensate for the small optical and mechanical misalignments of even the most precise telescope mounts, a mathematical model known as the pointing model is used to bring the initial pointing errors down to less than a few arc seconds. STELLA uses the classical pointing model, which describes only effects of misalignment of an otherwise perfect mount. The corrections to the azimuth () and the elevation (), namely, and , are modeled according to (3). and are constant offsets; the two angles and describe the tilt of the telescope's azimuth axis to the true vertical in the northern and eastern direction, respectively. describes the nonperpendicularity between the telescope's altitude and azimuth axis. is the nonperpendicularity between the telescope's altitude axis and the optical axis, while is the tube flexure

Ignoring the constant offset and , the highest absolute values are for and for and at and , but note also the discussion on pointing models in [12]. One specific effect of the mount of the STELLA telescopes is depicted in Figure 4; the tripod-mooring of the telescope mount has its resemblance in a ,wavelength modulated RMS in the elevation pointing corrections.

3.3. Focusing with a Focus Pyramid

Different solutions for focusing a telescope exist. On Stella-I, we tried to aim at a fast and reliable procedure. We decided to equip the telescope with the so-called focus pyramid, which can be rotated into the optical beam. It splits the light of a point-source, that is, of stars, into four individual subimages, whose distances are a direct measure of the focus; see Figure 5. Once calibrated, a single measurement of the subimages’ distances suffices to determine the focus.

3.4. Acquiring the Target

Currently, Stella-I is fiber-feeding an Echelle spectrograph [3]. Thus target acquisition must be done at a precision of less than an arc second. During integration, an adapter unit allows permanent position control down to subarc second levels. This adapter unit hosts a gray beam splitter that diverts 4% of the target light onto a guiding camera. The region just around the fiber entrance is imaged with a mirror onto the same guiding camera, leading to a second image at a varying offset from the direct beam splitter image; see Figure 6(a). The guiding camera has a shutter-less design to minimize the number of moving parts in the system. The read-out strips inherent to such a design are dealt with in the acquisition software. For details of the optical layout of the acquire unit, please refer to [1].

Once the telescope has been moved to the target position, an image of the entire guider's field of view (2.11.5 arcmin) is taken; see Figure 6(a). The image is bias-subtracted and a truncated Gaussian filter is applied to it. Similar to DAOfind [13], the resulting image is used to detect star candidates. Their sharpness and their elongation are validated against a two-dimensional probability function, shown in Figure 6(b). If no stars are found, the initial acquire is repeated up to five times with gradually increased exposure times. Once an ensemble of stars has been identified, their positions are matched to the UCAC2 [14] catalog and the shift to the current telescope's position is determined. After the first successful shift, this procedure is repeated using a much smaller window centered around the mirrored fiber entrance to finally reach an offset of less than . This relatively large offset is necessary because image motion induced by atmospheric turbulence at short exposure times can amount to an RMS in the target position of up to 0.5′′.

3.5. Closed-Loop Guiding on the Target Star

Once the star has been successfully acquired, the closed-loop guiding system starts. Its only aim is to keep the direct image of the target star as close to the mirrored fiber entrance position as possible. The target star's brightness defines the exposure time during the guiding phase; it is adjusted to get a stellar signal at a ratio of . The read-out time of the guider window limits the exposure time to 500 milliseconds. Typically, 3000 guider frames are taken during a one-hour integration. A combined frame of all individual guider frames can be seen in Figure 7. This combined frame is also used to measure the light loss at the fiber entrance by aperture photometry of the two stellar images.

Due to atmospheric image motion, wind shake, and intrinsic telescope oscillations, not every offset measured should be applied directly to the telescope. In STELLA, a split approach is used. First, five offsets are averaged. If the average offset is less than the standard deviation, or if it is less than the expected image motion induced by the atmosphere, it is set to zero. This average offset is split into azimuth and altitude and is fed into two PID controls (from Proportional-Integral-Derivative; for an introduction to PID refer to [15]). The output of the PIDs is then applied to the telescope. During commissioning, three distinct weather situations have been identified, each triggering the use of a different parameter set. In normal mode, a proportional term of for altitude and for azimuth is used. The integral constant I equals , and for the derivative term D, and are used. The bad-seeing mode, effective when the seeing is worse than 1.5′′, has a reduced , and no D or I term. In high-wind mode for wind speeds  m/s, the number of individual offsets averaged increases from five to 20, therefore acting on a much slower time scale. is thus relatively high, , , and a low I term of is introduced. stays at zero.

4. The Imaging Telescope STELLA-II

Different to the spectroscopic telescope, commissioning on the imaging telescope is just beginning. What follows are first results and insights gained, but far from being backed by years-long experience as in the spectroscopic case. Nevertheless, we want to address a few problems and their possible solutions in high-precision robotic photometry. For a more detailed description on the capabilities of the Wide Field STELLA Imaging Photometer (WiFSIP), see, for example, [16].

4.1. Ensuring High-Quality Flat Fielding

We do not intend to equip the STELLA building with a flat-fielding screen. The time of twilights will be used to obtain sky flats. The total number of 21 filters available in WiFSIP makes it impossible to obtain twilight flats on all filters each night. The individual filter will be grouped together according to their filter system (Sloan, Strømgren, Johnson-Cousins), and flats in single, but entire groups should be obtained in a single twilight. Two main drivers will define which group of flats will be chosen, the time passed since the last calibration of the filter group, and the likeliness of science observations with filters of that group in the upcoming night.

Following the text book, we start with the filter of the shortest central wavelength on dusk twilight and with the red filter on dawn twilight. Small-window test exposures are taken in rapid succession until the average ADU level in the read-out window suggests an exposure time in the allowed range −0.2 to 5 seconds. The Bonn shutter of the instrument allows for such very short exposures. On OmegaCam, a similar shutter is reported by [17] to deliver equally exposed images at exposure times down to 0.1 second.

To minimize the effect of sky-brightness gradients, the telescope is pointed to the anti solar azimuth and five degrees away from the zenith. This region in the sky is the one with the lowest, sometimes even vanishing sky brightness gradient; see [18]. Additionally, the camera will be rotated by after half of the individual exposures. Exposure times are gradually changed to keep the average exposure level as constant as possible.

The quality of the flat-field image is assessed via ADU statistics, with particular focus on higher-order moments. In Figure 8, the kurtosis of the ADUs for five sky flats, taken in Johnson-Cousins BVRI at two different days, is plotted. The flats are from the Mont Sec robotic observatory [19], separated by five days. Though the average ADU level dropped from almost 50000 to below 30000, the kurtosis, or, equivalent the fourth central moment, stayed remarkably constant. Cosmics show up in the kurtosis as high values. Higher moments then the kurtosis works even better in detecting cosmics but also probes the arithmetic accuracy of the CPU.

4.2. Focusing Using Statistical Moments

On the imaging telescope, no auxiliary focusing unit like a focus pyramid will be present. The plan is to use a twofold approach. Depending on the temperature of the telescope structure, a focus position will be estimated. At this estimated focus position together with offsets at 0.05 mm, three images of a field close to the celestial North pole (or maybe of Polaris itself, this will be decided during on-site commissioning) will be taken, and the kurtosis of the images will be analyzed. Figure 9 shows the anticipated procedure; the focus will be found at the point maximizing the kurtosis of the image. We hope that this procedure is superior to directly minimizing the FWHM of stellar images.

5. Future Plans

In spring 2010, the spectrograph-feeding fiber will be moved from Stella-I to the second STELLA telescope. The light coupling occurs at F1, leaving no place for a large acquire field. Acquiring will be done using a piggy-back auxiliary telescope, a 15 cm refracting telescope equipped with a shutter-less, uncooled guiding camera. Again, details on the optical layout can be found in [2].

5.1. Guiding on Spilled-over Light

The main acquiring and guiding system will be built around the current acquire and guiding logic, but with one additional complexity. Due to unavoidable, relative bending of the auxiliary telescope to the main telescope, guiding with the auxiliary telescope alone will not be sufficient. Additionally, we will use the light spilled over on the fiber entrance as a secondary guiding signal. This light will be caught on a fast read-out Unibrain 520b video camera. This camera comes with an electronic shutter, allowing for exposure times lower than 1ms and up to 65 seconds. Furthermore, the gain of the CCD can be adjusted quasicontinuously from 2.8 to 0.11. This extreme range will suffice to guide on stars as bright as zeroth magnitude down to, say, 12. First tests of the camera on Robotel allowed to estimate the quantum efficiency (QE) of the entire system to QE 0.15. Using artificial images, see Figure 10, we tested three algorithms to recover the shift of the stellar image with respect to the center of the pinhole: Center of gravity (CoG, based on image moments), quadrant weights (QWs), where the total ADUs in the left-to-right and up-to-down sectors have been compared to give a correction direction, and direct modelling of the star plus pinhole. Only direct modelling returns a true shift; CoG and QW can only deliver the principal direction of the correction. As guiding is done in a PID environment anyhow, this shortcoming does hardly matter as it is compensated by a proper choice of the proportional term.

The results of all three methods are shown in Figure 10(b). Basically all of the three methods could be used to retrieve the original shift. The higher robustness favors CoG and QW over direct modelling, while the latter delivers as a side product also information on light loss and stellar FWHM. Consequentially, we will use CoG during the guiding process and model the star plus pinhole on the combined frame to extract the additional information on light loss and seeing.