- Apr 29, 2014
-
-
Tomasz Grabiec authored
The test verifies basic tracing functionallity: * that samples are collected by OSv * extraction via `trace extract` succeeds * `trace summary` shows expected tracepoint names * listing via `trace list` shows tracepoints with arguments * network packet parsing works This test is run as part of standard 'make check'. Note: this single test can be executed as easily as: $ scripts/test.py --test tracing_smoke_test Signed-off-by:
Tomasz Grabiec <tgrabiec@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Tomasz Grabiec authored
Since we adveritse use of a system package in README, we should also recommend the same here. Signed-off-by:
Tomasz Grabiec <tgrabiec@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Tomasz Grabiec authored
It's needed by TCP-related tracing commands, as of now: trace.py list --tcpdump trace.py tcpdump Signed-off-by:
Tomasz Grabiec <tgrabiec@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Calle Wilund authored
Also, move platform dependent fast dispatch to platform arch code tree(s) The patching code is a bit more complex than would seem immediately (or even factually) neccesary. However, depending on cpu, there might be issues with trying to code patch across cache lines (unaligned). To be safe, we do it with the old 16-bit jmp + write + finish dance. [avi: fix up build.mk] Signed-off-by:
Calle Wilund <calle@cloudius-systems.com> Signed-off-by:
Avi Kivity <avi@cloudius-systems.com>
-
- Apr 28, 2014
-
-
Avi Kivity authored
Signed-off-by:
Avi Kivity <avi@cloudius-systems.com>
-
Tomasz Grabiec authored
tcpdump port number conversions vary from system to system. Turn of the conversions to have consistent output which is also simpler to assert for in tests. Signed-off-by:
Tomasz Grabiec <tgrabiec@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Tomasz Grabiec authored
On python 2.7.5 python complains about struct() not accepting unicode argument. Fix by adding explicit conversion. Signed-off-by:
Tomasz Grabiec <tgrabiec@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Avi Kivity authored
Signed-off-by:
Avi Kivity <avi@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Avi Kivity authored
phys_ptr<T>: unique_ptr<> for physical memory make_phys_ptr(): allocate and initialize a phys_ptr<> make_phys_array(): allocate a phys_ptr<> referencing an array Signed-off-by:
Avi Kivity <avi@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Gleb Natapov authored
Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Gleb Natapov authored
If dentry has elevated reference count during unlink (a file is opened for instance) it is not destroyed and remains in the dentry hash. As a result unlinked file is still accessible. Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Gleb Natapov authored
Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Gleb Natapov authored
If dentry has elevated reference count during rename (a file is opened for instance) it is not destroyed and remains in the dentry hash. As a result old file is accessible by destination path instead of new one. Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Gleb Natapov authored
Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Gleb Natapov authored
If dentry has elevated reference count during rename (a file is opened for instance) it is not destroyed and still hashed with old path. As a result renamed file can still be accessed by its old name. Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Takuya ASADA authored
Move static variables for alarm and related function to itimer class. itimer class offers interfaces to implement setitimer/getitimer, so added setitimer/getitimer. Now alarm rewrite as a just a wrapper function to call setitimer. Signed-off-by:
Takuya ASADA <syuu@cloudius-systems.com> Signed-off-by:
Avi Kivity <avi@cloudius-systems.com>
-
Takuya ASADA authored
Signed-off-by:
Takuya ASADA <syuu@cloudius-systems.com> Signed-off-by:
Avi Kivity <avi@cloudius-systems.com>
-
Nadav Har'El authored
When a small allocation is requested with large alignment, we ignored the alignment, and as a consequence posix_memalign() or alloc_phys_contiguous_aligned() could crash when it failed to achieve the desired alignment. This is not a common case (usually, size >= alignment, and the new C11 aligned_alloc() even supports only this case), but still it might happen, and we saw it in cloudius-systems/capstan#75. When size < alignment, this patch changes the size so we can achieve the desired alignment. For small alignments, this means setting size=alignment, so for example to get an alignment of 1024 bytes we need at least 1024-byte allocation. This is a waste of memory, but as these allocations are rare, we expect this to be acceptable. For large alignments, e.g., alignment=8192, we don't need size=alignment but we do need size to be large enough so we'll use malloc_large() (malloc_large() already supports arbitrarily large alignments). This patch also adds test cases to tst-align.so to test alignments larger than the desired size. Fixes #271 and cloudius-systems/capstan#75. Signed-off-by:
Nadav Har'El <nyh@cloudius-systems.com> Signed-off-by:
Avi Kivity <avi@cloudius-systems.com>
-
Asias He authored
AHCI spec says cmd_table need to be aligned to 128 bytes, respect it. AHCI spec does not specify the alignment requirement of _cmd_list and _recv_fis. To be safe, we align to their size, i.e. 1024 and 256 bytes. This fixes the issue reported by Roman Shaposhnik. $ capstan run -p vbox Created instance: i1398297574 OSv v0.07 Assertion failed: !(reinterpret_cast<uintptr_t>(ret) & (align - 1)) (/home/penberg/osv/core/mempool.cc: alloc_phys_contiguous_aligned: 1378) [backtrace] 0x398316 <memory::alloc_phys_contiguous_aligned(unsigned long, unsigned long)+214> 0x3542bb <ahci::port::setup()+283> 0x3549ae <ahci::port::port(unsigned int, ahci::hba*)+142> 0x354a5f <ahci::hba::scan()+111> 0x354d34 <ahci::hba::hba(pci::device&)+148> 0x354e07 <ahci::hba::probe(hw::hw_device*)+119> 0x33e65c <hw::driver_manager::register_driver(std::function<hw::hw_driver* (hw::hw_device*)>)+188> 0x33c8c3 <hw::device_manager::for_each_device(std::function<void (hw::hw_device*)>)+51> 0x33e3de <hw::driver_manager::load_all()+78> 0x20edf5 <do_main_thread(void*)+629> 0x3ff0a5 <sync+69> 0x3a3eaa <thread_main_c+26> 0x361585 <thread_main+7> Signed-off-by:
Asias He <asias@cloudius-systems.com> Signed-off-by:
Avi Kivity <avi@cloudius-systems.com>
-
- Apr 27, 2014
-
-
Avi Kivity authored
Signed-off-by:
Avi Kivity <avi@cloudius-systems.com>
-
- Apr 25, 2014
-
-
Dmitry Fleytman authored
There are 2 types of HVM AMIs on Amazon: 1. Windows based that support all instances types 2. Linux based that support instances larger than medium only Linux based AMIs suit OSv better because their default seciurity policy exactly fits OSv model (SSH access) and doesn't require reconfiguration. Windows based AMIs require default security policy reconfiguration in order to allow SSH access but from other hand allow running OSv on free-tier eligible micro instances which is good for enthusiasts and initial evaluations. In order to support both options this patch extends release scripts with ability to choose between Windows and Linux based templates during AMIs creation. Signed-off-by:
Dmitry Fleytman <dmitry@daynix.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Pekka Enberg authored
This reverts commit 532abb89 which makes Mr. Jenkins unhappy. Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Tomasz Grabiec authored
The test verifies that: * samples are collected by OSv * extraction via `trace extract` succeeds * `trace summary` shows expected tracepoint names * listing via `trace list` shows expected tracepoints * network packet parsing works This test is run as part of standard 'make check'. Note: this single test can be executed as easily as: $ scripts/test.py --test tracing_smoke_test Signed-off-by:
Tomasz Grabiec <tgrabiec@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Tomasz Grabiec authored
The command runs GDB which is very chatty. This will spoil our clean output of 'make check' once the tracing test is added. Let's hide GDB output unless it fails or the tracefile is not created. Signed-off-by:
Tomasz Grabiec <tgrabiec@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Tomasz Grabiec authored
The --tcpdump switch enables in-line decoding of net_packet* samples so that packet content is displayed in human-readable form next to other samples. This allows to corelate packets with other samples easily. Signed-off-by:
Tomasz Grabiec <tgrabiec@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Tomasz Grabiec authored
The 'scripts/trace.py pcap-dump' command will process all packet capturing samples and output a pcap stream. This is the format used by tcpdump. It can be used together with tcpdump to display packets in human-readable form like this: scripts/trace.py pcap-dump | tcpdump -r - | less Because pcap stream is often used together with tcpdump, I also introduce a shorthand for the above: scripts/trace.py tcpdump Signed-off-by:
Tomasz Grabiec <tgrabiec@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Tomasz Grabiec authored
There was no way to sniff packets going through OSv's loopback interface. I faced a need to debug in-guest TCP traffic. Packets are logged using tracing infrastructure. Packet data is serialized as sample data up to a limit, which is currently hardcoded to 128 bytes. To enable capturing of packets just enable tracepoints named: - net_packet_loopback - net_packet_eth Raw data can be seen in `trace list` output. Better presentation methods will be added in the following patches. This may also become useful when debugging network problems in the cloud, as we have no ability to run tcpdump on the host there. Signed-off-by:
Tomasz Grabiec <tgrabiec@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Tomasz Grabiec authored
Tracepoint argument which extends 'blob_tag' will be interpreted as a range of byte-sized values. Storage required to serialize such object is proportional to its size. I need it to implement storage-fiendly packet capturing using tracing layer. It could be also used to capture variable length strings. Current limit (50 chars) is too short for some paths passed to vfs calls. With variable-length encoding, we could set a more generous limit. Signed-off-by:
Tomasz Grabiec <tgrabiec@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Pekka Enberg authored
-
Avi Kivity authored
If the rcu threads need memory, let them have it, since they will use it to free even more memory. Reviewed-by:
Glauber Costa <glommer@cloudius-systems.com> Signed-off-by:
Avi Kivity <avi@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Avi Kivity authored
malloc() must wait for memory, and since page table operations can allocate memory, it must be able to dip into the reserve pool. free() should indicate it is a reclaimer. Reviewed-by:
Glauber Costa <glommer@cloudius-systems.com> Signed-off-by:
Avi Kivity <avi@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Avi Kivity authored
We already have a facility that to indicate that a thread is a reclaimer and should be allowed to allocate reserve memory (since that memory will be used to free memory). Extend it to allow indicating that a particular code section is used to free memory, not the entire thread. Reviewed-by:
Glauber Costa <glommer@cloudius-systems.com> Signed-off-by:
Avi Kivity <avi@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Nadav Har'El authored
After the previous patches, when we try to run an executable we cannot read (e.g., a directory - see issue #94), a "struct error" exception will be thrown out of osv::run, and nobody will catch it so the user will see a somewhat-puzzling "uncaught exception" error. With this patch, we catch the read error exception inside osv::run(), and when it happens, just return a normal load failure (nullptr). E.g, now trying to run a directory will result in a normal failure: $ scripts/run.py -e / OSv v0.07-39-g03feb99 run_main(): cannot execute /. Powering off. Fixes #94. The osv::run() API currently (before this patch, and also after it) doesn't have any way to say *why* the loading failed - it could have been that the executable was a directory, that it was not an ELF shared object, that it was a shared object and didn't have a main - in all cases the return value is nullptr. In the future this should probably change. Signed-off-by:
Nadav Har'El <nyh@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Nadav Har'El authored
read(fileref, ...) and write(fileref, ...), when they notice an error, currently assert() causing a crash which looks to non-expert users like an OSv bug. This causes, for example, bug #94, where doing osv::run() of a directory, instead of an executable, causes such an assertion failure. With this patch, these read() and write() will throw an exception (a struct error) instead of asserting. This fixes #94, as now the user will see an uncaught exception instead of buggy-looking assertion failure. In the next patch we'll catch this exception, so the user won't even see that. Signed-off-by:
Nadav Har'El <nyh@cloudius-systems.com> Signed-off-by:
Pekka Enberg <penberg@cloudius-systems.com>
-
Gleb Natapov authored
No longer used. Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com>
-
Gleb Natapov authored
The bug that caused it to be disabled should be fixed now. Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com>
-
Gleb Natapov authored
Now vma_list_mutex is used to protect against races between ARC buffer mapping by MMU and eviction by ZFS. The problem is that MMU code calls into ZFS with vma_list_mutex held, so on that path all ZFS related locks are taken after vma_list_mutex. An attempt to acquire vma_list_mutex during ARC buffer eviction, while many of the same ZFS locks are already held, causes deadlock. It was solved by using trylock() and skipping an eviction if vma_list_mutex cannot be acquired, but it appears that some mmapped buffers are destroyed not during eviction, but after writeback and this destruction cannot be delayed. It calls for locking scheme redesign. This patch introduce arc_lock that have to be held during access to read_cache. It prevents simultaneous eviction and mapping. arc_lock should be the most inner lock held on any code path. Code is change to adhere to this rule. For that the patch replaces ARC_SHARED_BUF flag by new b_mmaped field. The reason is that access to b_flags field is guarded by hash_lock and it is impossible to guaranty same order between hash_lock and arc_lock on all code paths. Dropping the need for hash_lock is a nice solution. Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com>
-
Gleb Natapov authored
Currently page_allocator return a page to a page mapper and the later populates a pte with it. Sometimes page allocation and pte population needs to be appear atomic though. For instance in case of a pagecache we want to prevent page eviction before pte is populated since page eviction clears pte, but if allocation and mapping is not atomic pte can be populated with stale data after eviction. With current approach very wide scoped lock is needed to guaranty atomicity. Moving pte population into page_allocator allows for much simpler locking. Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com>
-
Gleb Natapov authored
Current code assumes that for the same file and same offset ZFS will always return same ARC buffer, but this appears to be not the case. ZFS may create new ARC buffer while an old one is undergoing writeback. It means that we need to track mapping between file/offset and mmapped ARC buffer by ourselves. It's exactly what this patch is about. It adds new kind of cached page that holds pointers to an ARC buffer and stores them in new read_cache map. Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com>
-
Gleb Natapov authored
Signed-off-by:
Gleb Natapov <gleb@cloudius-systems.com>
-