[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
RE: [openrisc] Cache Line Fill
Hi Damjan,
Please let me know if you already checked in the new data base for "I
RAM replacement for I cache".
I hope you did have a chance to do the final testing.
Thanks
Michael Phan
----- Original Message -----
From: "Michael Phan" <mphan@n... >
To: <openrisc@o... >
Date: Thu, 26 Jun 2003 14:41:17 -0700
Subject: RE: [openrisc] Cache Line Fill
>
>
> Sound good to me.
>
> Thanks
>
> Michael Phan
>
>
> -----Original Message-----
> From: owner-openrisc@o...
> [mailto:owner-openrisc@o... ]On Behalf Of Damjan Lampret
> Sent: Thursday, June 26, 2003 11:37 PM
> To: Openrisc@O...
> Subject: Re: [openrisc] Cache Line Fill
>
>
> Hi Michael,
>
> yes I'm back. Thanks.
>
> I'm doing some final testing so I'm sure I don't check-in some
> garbage. I'll
> check-in tomorrow evening or saturday (depending when I'm done
with
> testing).
>
> regards,
> Damjan
>
> ----- Original Message -----
> From: <mphan@n... >
> To: <lampret@o... >; <openrisc@o... >
> Sent: Thursday, June 26, 2003 11:51 AM
> Subject: Re: [openrisc] Cache Line Fill
>
>
> > Hi Damjan,
> >
> > I hope you are back from travelling and had a nice one.
> > Did you check in the changes for "I RAM replacement for I
> cache" yet?
> > I am waiting to test it out.
> >
> > Thanks
> > Michael Phan
> >
> >
> > ----- Original Message -----
> > From: "Damjan Lampret" <lampret@o... >
> > To: <mphan@n... >,
> > <openrisc@o... >
> > Date: Sat, 7 Jun 2003 00:47:12 -0700
> > Subject: Re: [openrisc] Cache Line Fill
> >
> > >
> > >
> > > Michael,
> > >
> > > I assume you mean the I RAM replacement for I cache? Not
> yet, for
> > > the moment
> > > you can use RAM connected to iwb and I'll commit the
> changes to the
> > > cvs next
> > > weekend (I'm travelling this weekend until next weekend).
> > >
> > > regards,
> > > Damjan
> > >
> > > ----- Original Message -----
> > > From: <mphan@n... >
> > > To: <mphan@n... >;
> > > <openrisc@o... >
> > > Sent: Friday, June 06, 2003 11:11 AM
> > > Subject: Re: [openrisc] Cache Line Fill
> > >
> > >
> > > >
> > > > HI Damjan,
> > > >
> > > > Just want to touch base with you on this project,
> "instruction
> > > execution
> > > > with 0 wait state", did yo have a chane to put the
> changes
> > > into the
> > > > CVS so we can download and try them out.
> > > >
> > > > Thanks in advance
> > > > Michael Phan
> > > >
> > > >
> > > > ----- Original Message -----
> > > > From: mphan@n...
> > > > To: lampret@o... , openrisc@o...
> > > > Date: Tue, 6 May 2003 17:57:54 -0100
> > > > Subject: Re: [openrisc] Cache Line Fill
> > > >
> > > > >
> > > > >
> > > > >
> > > > > Hi Damjan,
> > > > >
> > > > > Your feedbacks are so precious and helpful to
> our
> > > project. Exactly,
> > > > > we
> > > > > don't need to use MMU and want to replace the
> cache with
> > > fixed
> > > > > memory for instruction execution with 0 wait
> state. So
> > > please put
> > > > > your
> > > > > changes in the CVS at your convenience so we
> can try it
> > > out and
> > > > > measure the performance improvement. Our
> project needs
> > > about 512
> > > > KB
> > > > > for the cache.
> > > > >
> > > > > Thousand thanks
> > > > > Michael Phan
> > > > >
> > > > >
> > > > >
> > > > > ----- Original Message -----
> > > > > From: "Damjan Lampret" <lampret@o... >
> > > > > To: <openrisc@o... >
> > > > > Date: Mon, 5 May 2003 21:21:38 -0700
> > > > > Subject: Re: [openrisc] Cache Line Fill
> > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > > ----- Original Message -----
> > > > > > From: <mphan@n... >
> > > > > > To: <openrisc@o... >
> > > > > > Sent: Thursday, May 01, 2003 3:36 PM
> > > > > > Subject: [openrisc] Cache Line Fill
> > > > > >
> > > > > >
> > > > > > > Hi Damjan.
> > > > > > >
> > > > > > > In the current design of orp_soc,
> with
> > > > > or1200_registered_ouput
> > > > > > > supported, a cache line fill takes 8
> clocks
> > > (2-2-2-2) to
> > > > > fetch
> > > > > > 4 DWORDs
> > > > > > > from the SRAM/FLASH. And a single
> DWORD fetch
> > > takes 3
> > > > > clocks
> > > > > > > (including one idle cycle of wb_cyc_o
> > > deasserted).
> > > > > > >
> > > > > > > If we have a very fast internal SRAM,
> is it
> > > possible to
> > > > > do a
> > > > > > cache line
> > > > > > fill
> > > > > > > with 4/5 clocks (1/2-1-1-1) by
> changing the
> > > wb_stb logic
> > > > > in
> > > > > > the
> > > > > > > or1200_wb_biu.v and do a single DWORD
> fetch
> > > with 2
> > > > > clocks.
> > > > > >
> > > > > > OR1200 is it is has been used in SoC
> projects much
> > > more
> > > > > complxed
> > > > > > than
> > > > > > orp_soc. In all these projects the memory
> subsystem
> > > takes more
> > > > > than
> > > > > > 1/2-1-1-1. So the current 2-2-2-2 was fast
> enough
> > > for all
> > > > > SoCs. If
> > > > > > you
> > > > > > however have faster memory subsystem than
> > > modification of
> > > > > > or1200_wb_biu and
> > > > > > possibly IC/DC state machines will be
> needed.
> > > > > >
> > > > > > >
> > > > > > > My next question is can we increase
> to cache
> > > size to 512
> > > > > kB to
> > > > > > reside
> > > > > > > the whole firmware and execute
> instructions
> > > from it with
> > > > > 0
> > > > > > wait state.
> > > > > > >
> > > > > >
> > > > > > If you want to use MMUs, then no. This is
> because
> > > MMU's page
> > > > > > translation is
> > > > > > done at the same time as cache access -
> virtual page
> > > number is
> > > > > > translated at
> > > > > > the same time as cache hit is determined.
> Since page
> > > size is
> > > > > 8KB,
> > > > > > largest
> > > > > > direct mapped cache can only be 8KB,
> unless you use
> > > several
> > > > > ways,
> > > > > > or unless
> > > > > > cache access takes an additional clock
> cycle (maybe
> > > acceptable
> > > > > for
> > > > > > data
> > > > > > accesses ?).
> > > > > >
> > > > > > Anyway if you don't need MMU, then your
> caches sizes
> > > are not
> > > > > > limited. To
> > > > > > increase cache size just add new IC/DC
> configuration
> > > (search
> > > > > for
> > > > > > "configuration" in or1200_defines.v and
> when you
> > > find IC and
> > > > > DC
> > > > > > configurations, just add a new size and
> then enable
> > > new
> > > > > > configuration).
> > > > > > Right now there are configurations for 4KB
> and 8KB
> > > caches.
> > > > > >
> > > > > > I'm working on one project where similar
> to your
> > > case all code
> > > > > > needs to be
> > > > > > accessible in 0 wait states. What I plan
> to do is to
> > > replaces
> > > > > > caches with
> > > > > > fixed memories - basically removing TAG
> RAMs and
> > > making sure
> > > > > that
> > > > > > the "hit"
> > > > > > always happens when accessing certain
> address range
> > > and never
> > > > > > happens when
> > > > > > accesssing outside of that range. This
> will
> > > effectively
> > > > > "change
> > > > > > caches into
> > > > > > fixed RAMs much like DSP RAMs or similar).
> > > > > > If you want these changes, I can put them
> into the
> > > cvs with
> > > > > > appropriate
> > > > > > defines. But it will take a few days.
> > > > > >
> > > > > > regards,
> > > > > > Damjan
> > > > > >
> > > > > > > Thanks
> > > > > > >
> > > > > > > Michael Phan
> > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > > >
> > >
> >
>
--
To unsubscribe from openrisc mailing list please visit http://www.opencores.org/mailinglists.shtml