From 280fa4d99e3d2dfbc3e8d4f006b62e4b78538677 Mon Sep 17 00:00:00 2001
From: David Li
Date: Mon, 28 Dec 2015 20:43:12 -0700
Subject: Specialize to direct-mapped cache
---
src/memory.rs | 56 +++++++++++++++++++++++++++++++++++---------------------
1 file changed, 35 insertions(+), 21 deletions(-)
(limited to 'src/memory.rs')
diff --git a/src/memory.rs b/src/memory.rs
index a8b15e2..c3cc60a 100644
--- a/src/memory.rs
+++ b/src/memory.rs
@@ -30,7 +30,7 @@ pub type Result = ::std::result::Result;
pub trait MemoryInterface {
fn latency() -> u32;
- fn read_word(&self, address: isa::Address) -> Result;
+ fn read_word(&mut self, address: isa::Address) -> Result;
fn write_word(&mut self, address: isa::Address, value: isa::Word) -> Result<()>;
// fn read_halfword(&self, address: isa::Address) -> Result;
@@ -63,17 +63,14 @@ struct CacheBlock {
fetch_request: Option,
}
-type CacheSet = Vec;
-
// TODO: probably want different caches for different strategies, and
// investigate how LRU is implemented
// TODO: use hashtable for a way?
// TODO: hashtable-based FA cache?
-pub struct Cache {
+pub struct DirectMappedCache {
num_sets: u32,
- num_ways: u32,
block_words: u32,
- cache: Vec,
+ cache: Vec,
next_level: T,
}
@@ -102,7 +99,7 @@ impl MemoryInterface for Memory {
100
}
- fn read_word(&self, address: isa::Address) -> Result {
+ fn read_word(&mut self, address: isa::Address) -> Result {
// memory is word-addressed but addresses are byte-addressed
self.memory.get((address / 4) as usize)
.map(Clone::clone)
@@ -122,17 +119,17 @@ impl MemoryInterface for Memory {
}
}
-impl Cache {
- pub fn new(sets: u32, ways: u32, block_words: u32, next_level: T) -> Cache {
- let set = vec![CacheBlock {
+impl DirectMappedCache {
+ pub fn new(sets: u32, block_words: u32, next_level: T)
+ -> DirectMappedCache {
+ let set = CacheBlock {
valid: false,
tag: 0,
contents: vec![0; block_words as usize],
fetch_request: None,
- }; ways as usize];
- Cache {
+ };
+ DirectMappedCache {
num_sets: sets,
- num_ways: ways,
block_words: block_words,
cache: vec![set; sets as usize],
next_level: next_level,
@@ -152,6 +149,11 @@ impl Cache {
(tag, index, offset)
}
+ fn normalize_address(&self, address: isa::Address) -> isa::Address {
+ let offset_mask = !(self.block_words * 4 - 1);
+ address & offset_mask
+ }
+
pub fn prefetch(&mut self, address: isa::Address) {
}
@@ -161,21 +163,33 @@ impl Cache {
}
}
-impl MemoryInterface for Cache {
+impl MemoryInterface for DirectMappedCache {
fn latency() -> u32 {
100
}
- fn read_word(&self, address: isa::Address) -> Result {
+ fn read_word(&mut self, address: isa::Address) -> Result {
+ let normalized = self.normalize_address(address);
let (tag, index, offset) = self.parse_address(address);
- let ref set = self.cache[index as usize];
- for way in set {
- if way.tag == tag {
- return Ok(way.contents[(offset / 4) as usize]);
- }
+ let ref mut set = self.cache[index as usize];
+ let stall = DirectMappedCache::::latency() + T::latency();
+ if set.tag == tag {
+ return Ok(set.contents[(offset / 4) as usize]);
+ }
+ else if let None = set.fetch_request {
+ set.fetch_request = Some(FetchRequest {
+ address: normalized,
+ prefetch: false,
+ cycles_left: stall,
+ })
+ }
+ else if let Some(ref fetch_request) = set.fetch_request {
+ return Err(MemoryError::CacheMiss {
+ stall_cycles: fetch_request.cycles_left
+ });
}
Err(MemoryError::CacheMiss {
- stall_cycles: Cache::::latency() + T::latency()
+ stall_cycles: stall,
})
}
--
cgit v1.2.3