TITLE: 09/14 Eve Shift: 2330-0500 UTC (1630-2200 PST), all times posted in UTC
STATE of H1: Corrective Maintenance
INCOMING OPERATOR: None
SHIFT SUMMARY:
IFO is in IDLE and CORRECTIVE MAINTENANCE
The "good" news:
DHARD P loop has been successfully automatic since closing it this afternoon. We can lock all the way to CARM_5_PICOMETERS quite quickly. DRMI has also been good to me this evening.
The bad news:
We thought we were losing lock at RESONANCE, but we were really losing lock at CARM_5_PICOMETERS.
Stepping through this state has showed that its only at the end of this, where the matrix involving REFL_BIAS, TR_REFL9 and TR_CARM is loaded, we lose lock roughly 10s later due to SRM becoming unstable quickly.
I investigated this via alogs 86909, 86910, 86911, 86912, which are comments to my OPS shift start (alog 86908). After being led down a rabbit hole of ramp times, from 3 other times where that was the problem, I can confirm that this isn't it. Curiously, lock was lost faster with a longer ramp time.
With confirmation from TJ and Jenne, I'm leaving IFO in IDLE with the plan being to solve this problem in tomorrow's shift.
I do feel like we're close to solving this problem or at least figuring out where the issue lies. Specific details are in the alogs listed below.
LOG:
None
TITLE: 09/13 Eve Shift: 2330-0500 UTC (1630-2200 PST), all times posted in UTC
STATE of H1: Corrective Maintenance
OUTGOING OPERATOR: Ryan S
CURRENT ENVIRONMENT:
SEI_ENV state: CALM
Wind: 17mph Gusts, 11mph 3min avg
Primary useism: 0.02 μm/s
Secondary useism: 0.15 μm/s
QUICK SUMMARY:
IFO is DOWN in CORRECTIVE MAINTENANCE
There was a lot of good work done today. Primarily rephasing REFLAIR (alog 86903) and closing DHARD_P loop. This has allowed us to easily get past DRMI. It seems that we are still losing lock around resonance due to a 35hz fast ringup.
The Procedure for the EVE:
What has happened so far:
Same PR_G ringup caused lockloss at RESONANCE and CARM_TO_REFL so I'm going to the state before, which is CARM_5_PICOMETERS. Will follow the blueprint above. Thankfully DHARD_P worked 2x now.
TITLE: 09/13 Day Shift: 1430-2330 UTC (0730-1630 PST), all times posted in UTC
STATE of H1: Corrective Maintenance
INCOMING OPERATOR: Ibrahim
SHIFT SUMMARY: The main efforts of today have been trying to get past DHARD_WFS. DRMI is now less troublesome since Sheila and Elenna rephased REFLAIR and some dark offsets were updated, and we are able to make it up to DARM_TO_RF consistently without intervention other than an occasional SRM alignment touchup. See the rephasing alog comments for details on locking efforts this afternoon.
LOG:
Start Time | System | Name | Location | Lazer_Haz | Task | Time End |
---|---|---|---|---|---|---|
18:39 | CDS | Richard | CER | N | Checking RF | 18:46 |
E. Capote, R. Short
While locking DRMI sevral times this morning, I had noticed the SRC loops are what have been pulling alignment away once DRMI ASC engages. I have put INP1, PRC1, and PRC2 back into Guardian to be turned on during DRMI ASC, and so far they have been working well. Elenna looked into this and noticed the dark offsets for sensors used for SRC needed to be updated, so she did so and the accepted SDFs are attached (all changes in the screenshot were accepted except for the last two in the list related to RF36).
Elenna, Ryan S, Sheila
This morning Ryan and Elenna were having difficulty getting DRMI to lock, so we locked PRMI and checked the phase of RELFAIR with the same template/PRM excitation described in 84630.
45 MHz phase changed by 7 degrees, 9MHz phase changed by 5 degrees. The attached screenshot shows that the phasing of RFL45 did have an impact on the OLG, and improved the phase margin. In that measurement we also added 25% to the MICH gain.
We accepted the phases in SDF so they were in effect at our next (quick) DRMI lock, but not the gain change.
When we next locked DRMI, we measured the MICH OLG again, and here it seems that the gain is a bit high. We haven't changed the gain in the guardian since this seemed to work well, but the third attached screen shot shows the loop gain.
After this we went to DARM to RF, and manually increased the TR_CARM offset to -7. The REFL power diped to 97% of it's unlocked value while the arm cavity transmission was 24 times the single arm level, so following the plot from 62110 we'd estimate our power recycling gain was between 20 and 30.
We made two attempts to close the DHAD WFS. In the first, we stepped the TR_CARM offset to -7 and tried the guardian steps. Looking at the lockloss it looked like the sign of both DHARD loops was wrong and pulling us to a bad alignment.
In the second attempt, we stayed at the TR_CARM offset of -3 (from DARM_TO_RF), and tried manually to engage them. The yaw loop was fine and we were able to use the normal full gain. For the pitch loop, it did seem to have the wrong sign so we tried flipping the sign. The guardian would step this gain from -1 to -10 at the end of the DHAR_WFS state, we stepped it from 1 to 3, which seemed to be driving the error signal to zero, but we lost lock partway through this.
I have made three more attempts through DHARD_WFS, both unsuccessful. Each time, once in DARM_TO_RF, I've manually engaged DHARD_Y's initial gain, watched the error signal converge, then increased to the final gain without issue. I then would engage DHARD_P's initial gain with the opposite sign and watch the error signal slowly converge after many minutes. Both times, whenever I would increase the DHARD_P gain, soon after the control signal would start moving away from zero (away from the direction it came) and there would be a lockloss.
On the thrid attempt I did the same as before, but this time stepped the CARM offset from -3 to -5 before engaging DHARD_P, but ultimately this attempt didn't work either. I noticed that once the DHARD_P error signal crosses zero, DHARD_Y starts running away (?). If I'm fast enough, I can turn the gains to zero before a lockloss happens, then bring the buildups back up by adjusting PRM. This juggling act of turning the DHARD loops on and off while adjusting PRM went on for a while, but inevitably ended in a lockloss.
Elenna, Ibrahim, Ryan, Sheila
We had one more attempt at locking, we were able to close DHARD Y wfs with the guardian, and we stepped DHARD P as we stepped up the CARM offset. Elenna fixed up DRMI along the way as well.
We engaged the DHARD P loop with the usual sign and gain when the TR_CARM offset was -35. Then we let the guardian continue, and lost lock in the CARM_TO_ANALOG state.
Ibrahim has a few plans of what to try next.
Sat Sep 13 10:08:37 2025 INFO: Fill completed in 8min 34secs
TITLE: 09/13 Day Shift: 1430-2330 UTC (0730-1630 PST), all times posted in UTC
STATE of H1: Corrective Maintenance
OUTGOING OPERATOR: None
CURRENT ENVIRONMENT:
SEI_ENV state: CALM
Wind: 4mph Gusts, 1mph 3min avg
Primary useism: 0.02 μm/s
Secondary useism: 0.17 μm/s
QUICK SUMMARY: I will start H1 running an alignment then try locking to see where we're still having trouble.
The SEI and SUS trips from the EQ moved alignment of the IMs quite a bit, so to lock XARM_IR, I restored these alignments based on top-mass OSEMs to what they were pre-trip. This was the only intervention needed for initial alignment. Starting main locking now.
Lockloss at DHARD_WFS. After locking DRMI, I tried moving PRM, SRM, and IM4 to get the buildups better, but I could not. I then was stepping through the following states slowly to see if an issue cropped up, and sure enough, after DHARD_WFS had completed I was waiting for DHARD to converge, and alignment was pulled away to eventually cause a lockloss; screenshot attached. I believe this was happening yesterday also. POP18/90 had been slowly degrading since locking DRMI, which may be because only MICH is enabled during DRMI ASC.
TITLE: 09/13 Eve Shift: 2330-0500 UTC (1630-2200 PST), all times posted in UTC
STATE of H1: Corrective Maintenance
INCOMING OPERATOR: NONE
SHIFT SUMMARY:
IFO is in IDLE for MAINTENANCE (and EQ) for the night. OWL is cancelled due to ongoing corrective maintenance from the most recent outage.
7.4 EQ from Kamchatka, Russia.
Sheila got to it first and untripped the tripped systems (alog 86896). I noticed that some of the suspensions (PRs and SRs) stalled so set them to their respective request states - and they got there successfully.
IFO is still in EQ mode but ringing down.
No HW WDs tripped and everything seems to be "normal" now.
LOG:
None
ops overvoew screenshot attached. I untripped things but will not try locking tonight.
While untripping suspensions I noticed that oddly IM2 watchdog rms monitors were just below the trip threshold, while the other IMs were 2-3 orders of magnitude smaller. The second screen shot shows this curious behavoir. What about IM2 undamped is so different from the other IMs?
Summary: omicron glitch rate increases after the power outage, and when the ISS loop is turned off
Using omicron, I made some plots of the glitch rates both before/after the power outage, and then again when the ISS loops were turn on/off. The times chosen were based on the detchar request made by Elenna in alog 86878. For the glitches, I used the omicron frequency/snr parameters of 10 < freq < 1000 Hz and 5 < snr < 500. The first slide in the attached pdf is a comparison of before/after the power outage. It's fairly obvous from the summary pages that the glitch rate increased, but I wanted to quantify just how much it changed. Before the power outage, it was around 45 glitches per hour, and after, it jumped up to about 600 glitches per hour. For both periods, it was around 10 hours of low noise I used for the comparison.
I then did a comparison between when the ISS loop was on versus off. I have found that the glitch rate increased when the ISS loop was turned back off, as can be seen in slide 2. On slide 3, I have summary page plots showing the WFS A YAW error signal, which looks noisier when the ISS loop is off, and the glitchgram below. On slide 4, I made a spectra comparing H1:CALIB_STRAIN_CLEAN during when the ISS loop was off/on, and we can see an increase in noise from ~ 0.06 Hz - 2 Hz, and at slightly higer frequencies of ~ 10 Hz - 40 Hz.
IFO is still trying to lock. I was given instructions to lock until PREP_ASC_FOR_FULL_IFO but after an initial alignment, the highest we could get to was CARM_TO_ANALOG. DRMI is just very unstable, even with manual intervention to maximize POP signal width.
IFO will keep trying to lock until PREP_ASC_FOR_FULL_IFO. I will change the request per instructions.
Writing this before I attempt to look at ISC_library but:
Our last stable state took us to CARM_5_PICOMETERS. I stayed there just to be sure it was stable.
I then followed the procedure and stepped through the next state, CARM_TO_REFL.
I will begin investigating what I can but I think this an issue with loading the matrix from ISC_Library, implying something is up with REFLBIAS or TR_REFL9? Somewhat out of my depth but will investigate alogs, wiki, ISC_library and anything else I can.
Either way, it seems that we thought we were losing at RESONANCE but it was the CARM_TO_REFL's last line, unless I got something wrong in the if-loop I didn't run through, which is also something I will investigate.
In looking at the problem lines loaded by the matrix, I see REFL_BIAS, TR_REFL9 and TR_CARM.
According to ISC_library.py, these are REFL_BIAS=9, TR_REFL9 = 27 and TR_CARM = 26. I don't know what this means yet.
Then, I looked at alogs hoping to find the last time something like this happened, which from vague memory was during post-vent recovery this year (with Craig and Georgia):
Looking for REFL_BIAS, I found Sheila's 2018 alog 43623 talking about the TR_CARM transition and how SRM would saturate first, which was what I was experiencing. Specifically the line is: TR_CARM transition: This has been unreliable for the last several days, our most frequent reason for failure in locking. We found that the problem was a large transient when the REFL_BIAS (TR CARM) path is first engaged would saturate SRM M1. We looked at the old version of the guardian and found that this used to be ramped on over 2 seconds, but has been ramped over 0.1 seconds recently. This transition has been sucsesfull both of the times we tried it since changing the ramp time, in one try there was no transient and in the other there was a transient but we stayed locked.
Looking for TR_REFL9, I found Craig's alog 84679 from the aforementioned post-vent recovery. Same issue with ramp time too, specifically referencing the few lines of ISC_LOCK that I posted above. He moved the ramp time to 2s, which I confirmed is still there. He has a picture of the lockloss (attached below) and it looks very similar to the ones we have been having. I trended the same channels he did and found that after this 2s ramp time, PR_GAIN would become unstable after 10s (attached below). Also found Elenna's alog (where I was also on-shift) dealing with the same issue, but before Craig had increased the ramp time - alog 84569. The deja vu was real. Buut I'm unsure this is the same issue because the ramp time is indeed 2s.
Looking for TR_CARM, I found this recent alog from Elenna discussing CARM, REFL and DHARD WFs - unknown whether it is relevant. Alog 86469. I will read this more closely.
While the I'm being led to just increase the ramp time further, I truly doubt this will change much since 2s is quite large of a ramp time and doesn't really explain much. Given all that, I'm going to try it before looking further since we've gotten back here in the time it took to write and investigate this.
As expected, that did not fix it, rather it seemed as though the faster ramp time caused a weird offset in REFLAIR9, which didn't happen last time. I checked the log and confirmed that there was a 4s instead of 2s ramp time but I'm kind of struggling to see it (attached).
There seems to be a 25hz signal on POP_A as soon as I increased the ramp time. Additionally, there wasn't really an instability upon turning on CARM_TO_REFL. It was stable for 9s until LL, showing a kick 3s before rather than a degradation. Additionally, the lockloss wasn't as violent.
With the tiniest font, maybe an even longer ramp time would work? Very unconvincing.
I don't have many ideas right now, but I'm reading the alog, the wiki and then devising some plan based off an idea of how this part of lock works.
I found an alog from Craig about REFLAIR9 (alog 43784). I'm somewhat more convinced that something involving the REFLAIR9_OFFSET and POP_LF?
The line from the alog: "we have made it through this transition every time now, but the PR_GAIN number on the striptool still spikes, meaning the POP_LF power changes rapidly through this transition, which is probably bad." sounds relevant?
I'm still confused why the REFLAIR9_OFFSET turning on causes an instability in PR_GAIN, which causes an instability in SRM which causes a lockloss.
I'm sufficiently confused at this stage so will attempt to go through the usual lockloss again to see if it's the same. Then I'll try stepping through again as a sanity check.
So far, I've only changed ramp times, immediately changing them back.
One thing that's confusing me is what the "REFL_BIAS" means. The REFL_BIAS Gain is 86 but the REFL_BIAS value in ISC_Library is 9.Nevermind, I think one refers to the location within a matrix in terms of the element number whereas the other is the value of that number within the matrix.