similar to: new features request

Displaying 20 results from an estimated 500 matches similar to: "new features request"

2004 Aug 06
2
new features request
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 On Fri, 13 Feb 2004, Geoff Shang wrote: > Mihai RUSU wrote: > > > Another nice feature to add would be a "bandwidth" limit (so just a new > > <limit> type entry). > > Problem with this is, what do you do once the limit has been reached? Who > do you kick? Or do you just not allow anymore listeners? I
2004 Aug 06
2
new features request
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 On Fri, 13 Feb 2004, Dave St John wrote: > > > > Another nice feature to add would be a "bandwidth" limit (so just a > new > > > > <limit> type entry). > If you are running icecast2 on linux you can handle bandwidth limitations > using QoS > more specificly iproute2. You mistunderstood me. No, I
2004 Aug 06
1
new features request
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 On Fri, 13 Feb 2004, Dave St John wrote: > I think that would be a good feature to add in icecast2, dont get me wrong, > im just saying with QoS > you have far much more flexibility in handling bandwidth. > tcp/ip packets can be marked with iptables and then ran threw tc which you > could then prioritize your packets You mean by using
2004 Aug 06
2
redundant code ?
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Hi In connection.c arround line 847 icecast2 does: if (global.clients >= client_limit) { client_send_504(client, "The server is already full. Try again later."); global_unlock(); return; } Then just a little bit later, does the same thing in the if (source) {} block. Is this code
2004 Aug 06
3
net/sock.c question
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Hi I have started to add bandwidth <limit> option to icecast. My ideea is to compute the current bandwidth by estimation on each sock_write*() call. Something like this: 1. initially we set kbps = 0, kbps_time = now and kbps_bytes = 0 2. after some time, a sock_write*() function is called which in turn calls sock_kbps_update(nobytes);
2004 Aug 06
2
icecast 2 auth problem with (old?) ices, MuSe
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Hi I have recently downloaded icecast 2 (I am an old icecast 1.x user) and tried to use it with my ices 0.2.3 (which works on icecast 1) but fails on login. Also I have tried with MuSe 0.8.1 and same error too. After some digging into the sources I found the problem to be with _check_pass_icy() which receives a "strange" pass to compare
2004 Aug 06
1
I need a Freelance Coder...
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Huh, why not just use ices with a perl song choser module ? This eliminates the need to develop the libshout based streamer (use ices instead) and give ices the title from your module, ices will take care of the rest. Or I am wrong ? :) On Wed, 25 Feb 2004, Renaud Waldura wrote: > For what's it's worth, I do something similar to this
2004 Aug 06
2
More libshout questions
Okay, I am using libshout to stream to a NullSoft Shoutcast server. I am using libmad and libmp3lame to do re-encoding on the fly. I am streaming to the server (currently) with shout_send_raw(). When I try to use shout_metadata_add(), I get a segfault in strdup(). Anyone have any ideas as to why? Everything else works fine, just not the metadata. Thanks In Advance, Deven --- >8 ----
2004 Aug 06
2
icecast 2 auth problem with (old?) ices, MuSe
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Hi Michael and thanks for answer! On Tue, 13 Jan 2004, Michael Smith wrote: > This shouldn't be being passed to _check_pass_icy() at all - this isn't the > "icy" (shoutcast nasty ugly) protocol at all. We support that (icy) protocol > because it's widely used - but it's a horribly ugly hack to pretend that
2004 Aug 06
0
new features request
> > > Another nice feature to add would be a "bandwidth" limit (so just a new > > > <limit> type entry). If you are running icecast2 on linux you can handle bandwidth limitations using QoS more specificly iproute2. That way icecast2 doesnt have to take up extra resources to handle bandwidth limitations. resources here cbq bash script
2004 Aug 06
0
new features request
I think that would be a good feature to add in icecast2, dont get me wrong, im just saying with QoS you have far much more flexibility in handling bandwidth. tcp/ip packets can be marked with iptables and then ran threw tc which you could then prioritize your packets set a limit and a burst for variable bitrates. (http://svana.org/kleptog/Packet-Shaping-HOWTO.txt) I know there isnt any "drop
2004 Aug 06
0
More libshout questions
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 On Fri, 16 Jan 2004, Deven Phillips wrote: > Okay, > > I am using libshout to stream to a NullSoft Shoutcast server. I am > using libmad and libmp3lame to do re-encoding on the fly. I am streaming > to the server (currently) with shout_send_raw(). When I try to use > shout_metadata_add(), I get a segfault in strdup(). Anyone have
2004 Aug 06
2
More libshout questions
Mihai, Here's the full backtrace. (gdb) bt full #0 0x40292093 in strlen () from /lib/i686/libc.so.6 No symbol table info available. #1 0x40291dc5 in strdup () from /lib/i686/libc.so.6 No symbol table info available. #2 0x400aed79 in _shout_util_dict_set (dict=0x0, key=0x804b746 "song", val=0x0) at util.c:236 prev = (util_dict *) 0x80555b8 #3 0x400ad61b in
2004 Aug 06
3
I need a Freelance Coder...
I must use Windows for Streaming. and the interface must be an port or something else, because the software interface runs on a dedicated server. ----- Original Message ----- From: "Ricardo Galli" <gallir@uib.es> To: <icecast-dev@xiph.org> Sent: Wednesday, February 25, 2004 10:10 AM Subject: Re: [icecast-dev] I need a Freelance Coder... <p>> On Wednesday 25 February
2008 Nov 18
2
pam_start() failed: system error
Hi, I'm using dovecot 1.1.3 on FreeBSD 6.3, x86, files are on NFS, except control and index files, which are local, on UFS2. It's moderately loaded, there are usually not more than 30 clients simultaneously. dovecot runs fine for 3-10 days, then people can't connect any more, and dovecot says: Nov 18 13:20:44 dizzy dovecot-auth: pam_ldap: ldap_simple_bind Can't contact LDAP
2004 Jun 21
3
[Fwd: Re: /sbin/tc does not know about HTB]
> It seems that the official place to grab the source for iproute2 is > http://developer.osdl.org/dev/iproute2/download/ but there are only > sources for 2.2 and 2.6, where are the ones for 2.4? Well I got mine from http://ftp.iasi.roedu.net/mirrors/ftp.inr.ac.ru/ip-routing/ but I didn''t even know about those 2.6 ones... I''m using the latest 2.4.7 with my 2.6.5
2015 Nov 18
3
[RFC PATCH 0/2] Google extension to improve qemu-nvme performance
Hi Rob & Mihai, I wrote vhost-nvme patches on top of Christoph's NVMe target. vhost-nvme still uses mmio. So the guest OS can run unmodified NVMe driver. But the tests I have done didn't show competitive performance compared to virtio-blk/virtio-scsi. The bottleneck is in mmio. Your nvme vendor extension patches reduces greatly the number of MMIO writes. So I'd like to push it
2015 Nov 18
3
[RFC PATCH 0/2] Google extension to improve qemu-nvme performance
Hi Rob & Mihai, I wrote vhost-nvme patches on top of Christoph's NVMe target. vhost-nvme still uses mmio. So the guest OS can run unmodified NVMe driver. But the tests I have done didn't show competitive performance compared to virtio-blk/virtio-scsi. The bottleneck is in mmio. Your nvme vendor extension patches reduces greatly the number of MMIO writes. So I'd like to push it
2003 Dec 29
5
include a file ?
ok, I've got yet another newbie question. My extensions.conf is getting rather longish and I'm getting dizzy moving back and forth editing this thing. Can I use the include command to include a file in order to break extensions.conf up into more manageable pieces ? Is breaking up the extension.conf file an OK thing to do ? Maybe something like this: include
2015 Nov 18
0
[PATCH -qemu] nvme: support Google vendor extension
From: Mihai Rusu <dizzy at google.com> This implements the device side for an NVMe vendor extension that reduces the number of MMIO writes which can result in a very large performance benefit in virtualized environments. See the following link for a description of the mechanism and the kernel NVMe driver changes to support this vendor extension: