Our Location
304 North Cardinal St.
Dorchester Center, MA 02124
For chassis with larger numbers of drives, or when connecting external JBOD chassis, it is common for the drives to connect to a specialized board that provides power and routing for the SATA/SAS signals to the controller. When building a storage system, there are many different ways the disks might be connected to the system. NVME-oF allows storage devices and arrays in remote chassis to be connected to local motherboards. NVMe storage comes in many form factors, from small M.2 devices to U.2 and other hot-swappable formats intended for servers. NVMe connects storage devices directly to the PCIe bus, offering extremely low latency and high throughput.
Unnamed devices can be specified by their specific SES device and element number. This greatly reduces the chance of getting it wrong when you (or the datacenter technician) physically pulls the disk. You can also reboot, and GEOM will pick up the multipath when it first tastes the disks during boot.
I moved my Scale server into the next room, laundry room, just so it’s out of sight. Replacing the drive is financially out of the question. I’m looking for a software solution, if possible, to make the HDD idling for most of the time when there is no load. Yeah, it’s not helping, thanks. Although it’s empty, so this is probably not the source of the constant HDD noise.
Most Seagate disks have configurable Extended Power Conditions (EPC) settings that include timers for how long the disk needs to stay idle before entering various low-power modes. Disk vendors typically provide their own vendor-specific ways to do persistent configuration of power management settings, so it’s worth trying to use those instead so the desired configuration doesn’t depend on the host system applying it, instead being configured in the drive (but in some cases it might be desirable to have the host configure that!). To prevent parking the heads at all a value greater than 128 may do the job (254 is a common choice, as the highest-power setting available), but it’s possible that some disks won’t behave this way because the ATA specification refers only to spinning down the disk and does not specify anything about parking heads. Typical SAS connectors support up to 4 drives per “lane”, but with an expander up to 255 devices are possible. An eight lane controller can only directly attach to 8 disks, requiring more controllers (consuming additional PCI-E slots) to connect more drives. This has long been the interface bus used by most home users to connect their hard drives, and is supported by nearly every motherboard.
In this case, there are at least two disks that I probably need to configure, since /dev/sde seems to be parking as often as about every 4 minutes (0.004 Hz) and /dev/sdc is only parking slightly less often. The smartmon_load_cycle_count_value metric seems like it would be the right one to query, but that actually expresses a percentage value (0-100) representing how many load cycles remain in the specified lifetime- on reaching 0 the disk has done a very large number of load cycles. It does support reading arbitrary metrics from text files written by other programs with its textfile collector however, which is fairly easy to integrate with arbitrary other tools. These communities are filled with knowledgeable individuals who can offer more personalized advice and help you navigate the complexities of long-term data storage.
Other interfaces for remote storage include iSCSI, Fiber-Channel, Infiniband, RoCE, and others, but those specialized solutions are beyond the scope of this article. Serial Attached SCSI (SAS) is the most common interface for enterprise storage, first appearing in 2004. Serial ATA (SATA) is the familiar interface used for non-enterprise storage, and is an extension of the original ATA interface dating from the 1980s. In this article we will discuss some strategies and tools to make managing disk arrays on FreeBSD (and related platforms like TrueNAS Core) much easier. It may be what you want is to enable HDD standby, which will “spin down” the drives when not in use
While I have been aware of this in my home server as well, it is easy to forget to ensure that disks are not silently killing themselves by cycling the heads. With modern, especially Enterprise grade hard drives being able to have hundreds of thousands of head park operations in their service life, is this really an isssue? With the tools presented here, the reader is well armed to react to failed disks and ensure that the wrong disk isn’t accidentally pulled. However, if a disk has died entirely, or a slot is empty, it might not have a device name. Sesutil can also be used to locate the disk in the physical array.While the SES data tells us that there is an 8 TB disk in Slot 06, it does not tell us which slot in the chassis corresponds to 06. Looking at a few items from the output, we can see the device names (/dev/da0 and /dev/da7 respectively) of the disks in Slot00 and Slot07.
At somewhat larger scales, a number of drives can be connected directly to a SAS (or SATA) controller PCIe card. But, if the number of ports on the motherboard is sufficient to your needs, this is the easiest way to connect the drives to the system. We are going to focus on some of the most popular for SATA and SAS drives.
I agree to receive your newsletters and accept the data privacy statement. Ensure device health & easy replacements with these valuable tips. Discover strategies to manage disk arrays on FreeBSD and related platforms/operating systems. Simply installing the apps and choosing a pool for k3s and docker creates a dataset and logs. Your pool gets writes from somewhere and ZFS is writing those to disk every 5 seconds.
I moved the system dataset to the boot pool. I don’t move any data, no apps are running, this is a vanilla Scale install so far, yet the HDD is in constant work. 1 SSD to boot and 1 HDD to store data. Agree, I have used SeaChest with good results for this same issue on scale plus drive cache. If you do it on a live pool, I’d back up your data first.
For ZFS users, automating fault responses with tools like reveryplay ZED (ZFS Event Daemon) can simplify disk replacement and minimize downtime. Configuring your system to notify you when a disk has errors, or when the filesystem reports a degraded device, will ensure your system gets prompt attention when something goes wrong. Experienced enterprise storage managers also keep extensive notes including the model number, SKU and/or URL for reordering, purchase order information, warranty end date, warranty URL, and any other useful information about each drive. While the operating system typically provides device aliases based on the disk’s serial number, WWN, or some other static identifier, this does not provide all of the information you might want.
SATA disks plugged directly into the motherboard use an interface called AHCI which does not provide much in the way of advanced management features. For smaller numbers of drives, and for most home systems, the most common way the disks are attached is to the SATA controllers built into the motherboard. Non-Volatile Memory Express (NVMe) is a newer storage interface that is becoming very popular for flash storage devices. Just download the executable file on both devices and run it to open the tool. At a glance, changing idle3 and EPC settings seems to have done the job nicely; here is the same graph of head park rates per disk as before, but on a smaller timescale that makes individual head parks visible. Seagate provide a “Seachest” collection of tools for manipulating their drives, but rather more usefully to users of non-Windows operating systems like Linux they also offer an open-source openSeaChest.
The settings you mentioned are already set this way. After you apply these settings the logs will be written to your SSD instead of being flushed to the disc array. Those are probably the system logs being flushed to disk every few seconds. I have moved the system data to my boot SSDs, don’t have any apps installed and don’t have any pool set for apps.
When it comes to long-term data storage, there are several strategies and media types that Redditors recommend. It refreshes the disks SMART information every 5 min. ZFS and Btrfs both aim to modernize storage by combining filesystems and volume management, but… Monitoring and maintaining your storage media is one of the most important parts of keeping your data safe.