Hi,
NickJohnson wrote:The user process can trash its own file handle table, but the important information relating to the file handles is stored within the respective drivers, safe from the user.
In this case, why does the user process need to keep its own file handle table, etc? For example, why can't a process just "exit()" without closing any file handles (or doing any other clean-up) if the device drivers need to be capable of doing the clean-up anyway.
NickJohnson wrote:And the user process can also send messages directly to the drivers, but the driver is notified of the caller's PID, so it can quickly filter requests, and certain messages are restricted to the kernel and drivers. The reason I said my message passing system is mostly DoS-proof is that messages (which are really more like *nix signals) can preempt each other, and sending a message ends a process' timeslice, so only so many message can be sent at once.
If messages can preempt each other, then you might end up with a re-entrancy nightmare. For example, a device driver receives a message and starts modifying a critical data structure, but is then pre-empted by a second message. To handle the second message the driver needs to use this critical data structure, but the critical data structure is in an inconsistant state because it hasn't finished modifying it. In this case if you use a re-entrancy lock, then you'll just get a dead-lock because after the driver is preempted it won't be able to acquire the lock a second time. There's no easy way to solve this problem.
One way to solve the problem would be to allow a device driver to prevent pre-emption while the driver is modifying the critical data structure; but then the kernel would need to wait (for an unknown length of time) before it can send a message to the driver.
Another way would be to use a special type of re-entrancy lock, where if the lock is already acquired you switch to the pre-empted code until it releases the lock, and then switch back. In this case you'd need to be careful of misuse (e.g. drivers that decide to switch to other tasks for other reasons).
NickJohnson wrote:Brendan wrote:If one process is a service that handles something that libsys doesn't know about, and a process that uses this service crashes, then who tells the service that the process crashed? For example, imagine if a developer writes their own font engine (the service), and then writes 3 different games that send requests to this font engine, and one of these games asks the font engine to load some fonts and then crashes without telling the font engine that it's finished. In this case, does the font engine end up with 20 MiB of font data that's never freed?
I think this is a real problem though. I guess the best solution would be to have a table in the process structure (in the kernel) that keeps track of which processes to notify when a client exits. Drivers could register themselves to be notified, but not regular processes.
That's how I do it - I call this "obituaries" (if any task is terminated for any reason, the any other task that has requested it is told about it).
If a device driver crashes, then how do regular processes know that any pending requests won't be honored?
For my OS design, if a device driver (or file system) crashes then the VFS gets the obituary, and because the VFS manages all file I/O it knows which requests are still pending and can return errors for any effected requests. For example, an application asks to read 1024 bytes from a file handle, the VFS checks it's cache and if the data isn't in the cache the VFS asks the file system driver to fetch the data, and if the file system driver crashes the VFS can return an error to the application. However, I also plan to use redundant file systems, where the VFS might be able to successfully complete the application's request by asking a different (mirrored) file system to fetch the data from the file.
Of course in my case, if the VFS crashes the OS is entirely screwed (but that's likely to be the case for all OSs, regardless of what you do).
I guess that in your case, you could allow regular processes (libsys) to register for obituaries and then have obituary handling code in libsys...
NickJohnson wrote:Brendan wrote:To improve performance, most OS's use the VFS to cache file data. For your OS will every file system need to maintain it's own cache? If the kernel is running out of RAM, will it need to send a message to every file system to ask it to free up some RAM (reduce the amount of data in it's cache); and how will each file system know if the data it's cacheing is more important (or less important) than the data that other file systems are cacheing (to make sure that the most important data is still cached and only the least important data is freed)?
My intention is to allow drivers to handle a lot of policy on this individually. I could make a system call that reads the current memory usage, so drivers could regulate their cache sizes.
If one file system has 5 MiB of cached data that hasn't been used for ages and another file system has 5 MiB of cached data that is used very often, which 5 MiB of data will be freed?
NickJohnson wrote:Brendan wrote:Most OS's also have something called "notifications"; where a process can ask the VFS to tell it if a certain file is changed.
Because of my system's purely event driven architecture, it would be easy to implement notifications within each driver as well. The VFS would have to synchronize with the state of the real filesystems, so it keeps track of mountpoints and symlinks internally (isn't that what the VFS does anyway?). I think the more interesting part is that you could have two or more VFSs set up with different mountpoints. E.g. VFS 1 could be *nix style (/mnt/device), but VFS 2 could be MS-DOS/Windows style (device:\).
Normally the VFS keeps track of everything, so that (for e.g.) if a process wants to read the file "/home/bcos/foo/bar/hello/goodbye.txt" the VFS knows exactly where this file is (including mountpoints and symbolic links) and knows if the process has permission to access the file, and if the data is already cached it can send the data to the process immediately (without the file system itself being involved).
If there's 2 different VFSs being used, and I create a symbolic link from "/home/bcos/mycd" to "/mnt/cdrom", then would the file system tell one VFS that there's a new symbolic link from "/home/bcos/mycd" to "/mnt/cdrom" and tell the other VFS that there's a new symbolic link from "c:\home\bcos\mycd" to "D:\"? Does this mean that all file systems need to know about all mount points for all VFSs?
I'd also point out that IMHO choice is good (e.g. allowing an arbitrary number of different VFSs), but standardization is a lot better (e.g. having one standard VFS that all users, administrators, technicians and programmers can become familiar with). If you've ever wondered why there's so little commercial software for Linux, just think about being a help desk operator for "XYZ accounting package" when a customer rings up because they've got problems installing it...
Cheers,
Brendan